var/home/core/zuul-output/0000755000175000017500000000000015116734760014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116762442015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006547714315116762434017727 0ustar rootrootDec 12 06:48:31 crc systemd[1]: Starting Kubernetes Kubelet... Dec 12 06:48:31 crc restorecon[4778]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:31 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 12 06:48:32 crc restorecon[4778]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 12 06:48:32 crc kubenswrapper[4867]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.647944 4867 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662558 4867 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662613 4867 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662624 4867 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662634 4867 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662643 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662651 4867 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662658 4867 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662664 4867 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662671 4867 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662678 4867 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662685 4867 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662691 4867 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662698 4867 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662705 4867 feature_gate.go:330] unrecognized feature gate: Example Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662712 4867 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662721 4867 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662729 4867 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662736 4867 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662744 4867 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662750 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662756 4867 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662765 4867 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662774 4867 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662783 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662792 4867 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662800 4867 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662807 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662815 4867 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662821 4867 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662826 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662831 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662837 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662853 4867 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662859 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662864 4867 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662871 4867 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662876 4867 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662882 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662887 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662893 4867 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662898 4867 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662903 4867 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662908 4867 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662915 4867 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662922 4867 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662928 4867 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662935 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662940 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662946 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662952 4867 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662957 4867 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662963 4867 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662968 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662973 4867 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662978 4867 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662984 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662990 4867 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.662995 4867 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663000 4867 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663005 4867 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663010 4867 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663015 4867 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663020 4867 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663026 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663030 4867 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663036 4867 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663041 4867 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663048 4867 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663053 4867 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663058 4867 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.663063 4867 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663434 4867 flags.go:64] FLAG: --address="0.0.0.0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663462 4867 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663478 4867 flags.go:64] FLAG: --anonymous-auth="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663487 4867 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663496 4867 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663503 4867 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663513 4867 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663521 4867 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663527 4867 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663534 4867 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663543 4867 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663553 4867 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663562 4867 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663570 4867 flags.go:64] FLAG: --cgroup-root="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663577 4867 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663585 4867 flags.go:64] FLAG: --client-ca-file="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663593 4867 flags.go:64] FLAG: --cloud-config="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663601 4867 flags.go:64] FLAG: --cloud-provider="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663608 4867 flags.go:64] FLAG: --cluster-dns="[]" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663618 4867 flags.go:64] FLAG: --cluster-domain="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663641 4867 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663648 4867 flags.go:64] FLAG: --config-dir="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663655 4867 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663661 4867 flags.go:64] FLAG: --container-log-max-files="5" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663670 4867 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663677 4867 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663684 4867 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663690 4867 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663697 4867 flags.go:64] FLAG: --contention-profiling="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663703 4867 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663709 4867 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663716 4867 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663722 4867 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663730 4867 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663737 4867 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663743 4867 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663749 4867 flags.go:64] FLAG: --enable-load-reader="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663755 4867 flags.go:64] FLAG: --enable-server="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663762 4867 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663918 4867 flags.go:64] FLAG: --event-burst="100" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663925 4867 flags.go:64] FLAG: --event-qps="50" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663932 4867 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663939 4867 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663946 4867 flags.go:64] FLAG: --eviction-hard="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663959 4867 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663966 4867 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663972 4867 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663980 4867 flags.go:64] FLAG: --eviction-soft="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663987 4867 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.663994 4867 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664000 4867 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664007 4867 flags.go:64] FLAG: --experimental-mounter-path="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664013 4867 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664019 4867 flags.go:64] FLAG: --fail-swap-on="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664025 4867 flags.go:64] FLAG: --feature-gates="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664041 4867 flags.go:64] FLAG: --file-check-frequency="20s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664050 4867 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664058 4867 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664066 4867 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664074 4867 flags.go:64] FLAG: --healthz-port="10248" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664081 4867 flags.go:64] FLAG: --help="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664089 4867 flags.go:64] FLAG: --hostname-override="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664097 4867 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664105 4867 flags.go:64] FLAG: --http-check-frequency="20s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664113 4867 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664121 4867 flags.go:64] FLAG: --image-credential-provider-config="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664129 4867 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664136 4867 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664143 4867 flags.go:64] FLAG: --image-service-endpoint="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664149 4867 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664156 4867 flags.go:64] FLAG: --kube-api-burst="100" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664162 4867 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664171 4867 flags.go:64] FLAG: --kube-api-qps="50" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664177 4867 flags.go:64] FLAG: --kube-reserved="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664184 4867 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664190 4867 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664196 4867 flags.go:64] FLAG: --kubelet-cgroups="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664203 4867 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664209 4867 flags.go:64] FLAG: --lock-file="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664216 4867 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664247 4867 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664255 4867 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664266 4867 flags.go:64] FLAG: --log-json-split-stream="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664275 4867 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664282 4867 flags.go:64] FLAG: --log-text-split-stream="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664288 4867 flags.go:64] FLAG: --logging-format="text" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664294 4867 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664301 4867 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664308 4867 flags.go:64] FLAG: --manifest-url="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664314 4867 flags.go:64] FLAG: --manifest-url-header="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664323 4867 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664330 4867 flags.go:64] FLAG: --max-open-files="1000000" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664338 4867 flags.go:64] FLAG: --max-pods="110" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664344 4867 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664352 4867 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664358 4867 flags.go:64] FLAG: --memory-manager-policy="None" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664365 4867 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664372 4867 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664378 4867 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664385 4867 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664401 4867 flags.go:64] FLAG: --node-status-max-images="50" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664408 4867 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664414 4867 flags.go:64] FLAG: --oom-score-adj="-999" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664421 4867 flags.go:64] FLAG: --pod-cidr="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664428 4867 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664437 4867 flags.go:64] FLAG: --pod-manifest-path="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664443 4867 flags.go:64] FLAG: --pod-max-pids="-1" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664450 4867 flags.go:64] FLAG: --pods-per-core="0" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664456 4867 flags.go:64] FLAG: --port="10250" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664463 4867 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664469 4867 flags.go:64] FLAG: --provider-id="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664475 4867 flags.go:64] FLAG: --qos-reserved="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664482 4867 flags.go:64] FLAG: --read-only-port="10255" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664488 4867 flags.go:64] FLAG: --register-node="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664496 4867 flags.go:64] FLAG: --register-schedulable="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664502 4867 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664514 4867 flags.go:64] FLAG: --registry-burst="10" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664521 4867 flags.go:64] FLAG: --registry-qps="5" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664527 4867 flags.go:64] FLAG: --reserved-cpus="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664534 4867 flags.go:64] FLAG: --reserved-memory="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664543 4867 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664549 4867 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664556 4867 flags.go:64] FLAG: --rotate-certificates="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664562 4867 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664568 4867 flags.go:64] FLAG: --runonce="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664575 4867 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664581 4867 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664588 4867 flags.go:64] FLAG: --seccomp-default="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664594 4867 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664600 4867 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664608 4867 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664614 4867 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664621 4867 flags.go:64] FLAG: --storage-driver-password="root" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664627 4867 flags.go:64] FLAG: --storage-driver-secure="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664634 4867 flags.go:64] FLAG: --storage-driver-table="stats" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664641 4867 flags.go:64] FLAG: --storage-driver-user="root" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664647 4867 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664654 4867 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664660 4867 flags.go:64] FLAG: --system-cgroups="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664666 4867 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664677 4867 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664684 4867 flags.go:64] FLAG: --tls-cert-file="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664690 4867 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664698 4867 flags.go:64] FLAG: --tls-min-version="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664704 4867 flags.go:64] FLAG: --tls-private-key-file="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664710 4867 flags.go:64] FLAG: --topology-manager-policy="none" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664717 4867 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664723 4867 flags.go:64] FLAG: --topology-manager-scope="container" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664730 4867 flags.go:64] FLAG: --v="2" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664738 4867 flags.go:64] FLAG: --version="false" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664747 4867 flags.go:64] FLAG: --vmodule="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664756 4867 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.664763 4867 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664915 4867 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664924 4867 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664931 4867 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664937 4867 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664944 4867 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664949 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664954 4867 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664960 4867 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664965 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664970 4867 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664975 4867 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664981 4867 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664986 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.664993 4867 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665000 4867 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665006 4867 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665012 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665018 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665025 4867 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665030 4867 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665036 4867 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665042 4867 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665048 4867 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665053 4867 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665059 4867 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665065 4867 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665070 4867 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665076 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665081 4867 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665087 4867 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665093 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665098 4867 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665104 4867 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665109 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665114 4867 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665119 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665125 4867 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665130 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665136 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665142 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665147 4867 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665153 4867 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665160 4867 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665185 4867 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665191 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665198 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665204 4867 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665210 4867 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665236 4867 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665242 4867 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665248 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665254 4867 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665259 4867 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665264 4867 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665270 4867 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665275 4867 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665280 4867 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665287 4867 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665292 4867 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665297 4867 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665303 4867 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665308 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665313 4867 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665319 4867 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665324 4867 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665329 4867 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665335 4867 feature_gate.go:330] unrecognized feature gate: Example Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665340 4867 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665345 4867 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665350 4867 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.665356 4867 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.665374 4867 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.672205 4867 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.672269 4867 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672347 4867 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672358 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672364 4867 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672372 4867 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672378 4867 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672383 4867 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672387 4867 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672391 4867 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672395 4867 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672399 4867 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672403 4867 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672407 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672410 4867 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672415 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672420 4867 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672424 4867 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672429 4867 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672614 4867 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672618 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672622 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672626 4867 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672630 4867 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672633 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672637 4867 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672641 4867 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672645 4867 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672648 4867 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672652 4867 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672655 4867 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672661 4867 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672665 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672669 4867 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672672 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672677 4867 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672682 4867 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672686 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672689 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672692 4867 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672696 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672700 4867 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672705 4867 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672709 4867 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672713 4867 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672718 4867 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672722 4867 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672726 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672730 4867 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672734 4867 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672738 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672741 4867 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672745 4867 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672748 4867 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672752 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672755 4867 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672759 4867 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672764 4867 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672767 4867 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672771 4867 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672776 4867 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672781 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672785 4867 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672788 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672792 4867 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672796 4867 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672800 4867 feature_gate.go:330] unrecognized feature gate: Example Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672809 4867 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672814 4867 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672818 4867 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672822 4867 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672825 4867 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672829 4867 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.672837 4867 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672957 4867 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672964 4867 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672968 4867 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672972 4867 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672977 4867 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672981 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672986 4867 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672990 4867 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672994 4867 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.672998 4867 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673001 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673006 4867 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673009 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673013 4867 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673016 4867 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673020 4867 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673025 4867 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673029 4867 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673033 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673037 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673042 4867 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673046 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673050 4867 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673055 4867 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673059 4867 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673063 4867 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673066 4867 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673070 4867 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673074 4867 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673080 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673084 4867 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673087 4867 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673091 4867 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673095 4867 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673098 4867 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673103 4867 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673106 4867 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673110 4867 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673114 4867 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673157 4867 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673163 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673167 4867 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673171 4867 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673175 4867 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673181 4867 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673185 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673189 4867 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673194 4867 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673198 4867 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673202 4867 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673206 4867 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673210 4867 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673213 4867 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673234 4867 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673238 4867 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673242 4867 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673246 4867 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673250 4867 feature_gate.go:330] unrecognized feature gate: Example Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673254 4867 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673258 4867 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673261 4867 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673265 4867 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673268 4867 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673272 4867 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673276 4867 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673281 4867 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673285 4867 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673289 4867 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673292 4867 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673296 4867 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.673300 4867 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.673307 4867 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.673525 4867 server.go:940] "Client rotation is on, will bootstrap in background" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676133 4867 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676240 4867 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676703 4867 server.go:997] "Starting client certificate rotation" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676726 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676906 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-04 02:46:43.584823687 +0000 UTC Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.676999 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.681353 4867 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.683027 4867 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.683057 4867 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.695450 4867 log.go:25] "Validated CRI v1 runtime API" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.709013 4867 log.go:25] "Validated CRI v1 image API" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.721197 4867 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.723345 4867 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-12-06-44-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.723390 4867 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.742632 4867 manager.go:217] Machine: {Timestamp:2025-12-12 06:48:32.73573748 +0000 UTC m=+0.307118739 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3195cc84-39c8-4f61-8d51-1e423683d247 BootID:b0f6ee30-6d15-417f-bdda-47a426b4f903 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:bb:a2:bf Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:bb:a2:bf Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:16:00:21 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3c:95:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:77:e1:72 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:69:b3:8d Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:94:58:9b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:84:e4:7a:da:7c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1e:23:ae:e7:5b:5b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.743157 4867 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.743355 4867 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.743685 4867 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.743872 4867 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.743913 4867 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.744160 4867 topology_manager.go:138] "Creating topology manager with none policy" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.744173 4867 container_manager_linux.go:303] "Creating device plugin manager" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.744400 4867 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.744444 4867 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.744714 4867 state_mem.go:36] "Initialized new in-memory state store" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745112 4867 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745708 4867 kubelet.go:418] "Attempting to sync node with API server" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745730 4867 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745753 4867 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745766 4867 kubelet.go:324] "Adding apiserver pod source" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.745781 4867 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.747517 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.747594 4867 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.747630 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.747815 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.747865 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.748113 4867 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.748999 4867 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749593 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749618 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749626 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749633 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749674 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749682 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749691 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749702 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749715 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749722 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749750 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749757 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.749930 4867 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.750545 4867 server.go:1280] "Started kubelet" Dec 12 06:48:32 crc systemd[1]: Started Kubernetes Kubelet. Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.751053 4867 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.760004 4867 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.760490 4867 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.760504 4867 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762536 4867 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762573 4867 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762597 4867 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 17:31:06.022632431 +0000 UTC Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762764 4867 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762777 4867 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.762908 4867 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.762868 4867 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.774388 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.774487 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.769697 4867 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.66:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18806501e597598a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-12 06:48:32.750508426 +0000 UTC m=+0.321889695,LastTimestamp:2025-12-12 06:48:32.750508426 +0000 UTC m=+0.321889695,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.774544 4867 server.go:460] "Adding debug handlers to kubelet server" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.774992 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="200ms" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.776900 4867 factory.go:55] Registering systemd factory Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.776932 4867 factory.go:221] Registration of the systemd container factory successfully Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.777413 4867 factory.go:153] Registering CRI-O factory Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.777455 4867 factory.go:221] Registration of the crio container factory successfully Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.777557 4867 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.777606 4867 factory.go:103] Registering Raw factory Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.777628 4867 manager.go:1196] Started watching for new ooms in manager Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.778328 4867 manager.go:319] Starting recovery of all containers Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792813 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792872 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792883 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792894 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792906 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792919 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792931 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792943 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792983 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.792999 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793015 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793031 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793043 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793060 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793071 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793083 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793094 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793105 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793116 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793126 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793137 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793174 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793185 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793196 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793245 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793257 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793270 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793284 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793296 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793309 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793339 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793352 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793362 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793374 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793386 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793398 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793409 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793421 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793431 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793442 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793453 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793463 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793477 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793486 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793496 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793506 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793517 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793528 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793540 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793550 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793560 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793571 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793585 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793597 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793609 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793619 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793628 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793643 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793655 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793664 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793673 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793682 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793690 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793701 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793709 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793718 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793727 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793737 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793748 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793756 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793796 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793806 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793816 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793831 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793841 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793851 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793861 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793878 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793893 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793907 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793919 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793937 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793948 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793957 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793967 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793977 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793986 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.793997 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794006 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794015 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794025 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794035 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794046 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794056 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794066 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794077 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794087 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794096 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794105 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794113 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794122 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794130 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794139 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794149 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794166 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794178 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794187 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794196 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794205 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794215 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794239 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794251 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794263 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794306 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794348 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794357 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794381 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794389 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794401 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794410 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794420 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794429 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794461 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794470 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794481 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794514 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794524 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794549 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794559 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794608 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794618 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794629 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794640 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794651 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794662 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794687 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794697 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794711 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794720 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794731 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794742 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794752 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794761 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794789 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794798 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794807 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794818 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794827 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794836 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794848 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794857 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794882 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794891 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794902 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794934 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.794990 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795000 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795011 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795022 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795046 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795055 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795065 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795073 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795083 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795106 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795115 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795128 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795185 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795196 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795206 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795216 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795276 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795290 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795302 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795315 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795329 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795340 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795352 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795389 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795400 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795429 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795445 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795455 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795466 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795476 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795488 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795525 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795540 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795553 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795565 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795576 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795611 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795622 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795632 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795681 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795701 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795738 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795749 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795796 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795833 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795845 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795857 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795869 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795882 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.795951 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.798846 4867 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.798893 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.798933 4867 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.798943 4867 reconstruct.go:97] "Volume reconstruction finished" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.798951 4867 reconciler.go:26] "Reconciler: start to sync state" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.804403 4867 manager.go:324] Recovery completed Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.826909 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.828844 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.828897 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.828907 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.829837 4867 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.829847 4867 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.829866 4867 state_mem.go:36] "Initialized new in-memory state store" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.834727 4867 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.836797 4867 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.836844 4867 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.836877 4867 kubelet.go:2335] "Starting kubelet main sync loop" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.836924 4867 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 12 06:48:32 crc kubenswrapper[4867]: W1212 06:48:32.837954 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.838026 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.839714 4867 policy_none.go:49] "None policy: Start" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.840695 4867 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.840724 4867 state_mem.go:35] "Initializing new in-memory state store" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.869798 4867 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.895407 4867 manager.go:334] "Starting Device Plugin manager" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.895487 4867 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.895502 4867 server.go:79] "Starting device plugin registration server" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.896208 4867 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.896247 4867 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.896417 4867 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.896534 4867 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.896542 4867 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.902406 4867 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.937644 4867 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.937808 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.939268 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.939312 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.939327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.939604 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.939988 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.940040 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.940594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.940626 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.940640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.940732 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.941275 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.941324 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.941272 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.941420 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.941435 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.942515 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.942548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.942564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943255 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943301 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943311 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943480 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943569 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.943608 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944276 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944309 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944320 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944555 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944736 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.944777 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945355 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945388 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945574 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945589 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945637 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945624 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945662 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945880 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.945921 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.946663 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.946690 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.946702 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.975954 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="400ms" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.997140 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.998610 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.998688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.998718 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:32 crc kubenswrapper[4867]: I1212 06:48:32.998761 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:32 crc kubenswrapper[4867]: E1212 06:48:32.999473 4867 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.66:6443: connect: connection refused" node="crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000584 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000626 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000665 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000697 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000733 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000769 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000829 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000925 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.000987 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001024 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001062 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001136 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001174 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001199 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.001244 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102677 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102745 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102765 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102785 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102805 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102843 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102869 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102890 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102916 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102936 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102956 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.102997 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103017 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103029 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103087 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103131 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103167 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103157 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103091 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103093 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103030 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103212 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103141 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103150 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103278 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103172 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103200 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103139 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.103028 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.199854 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.201213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.201290 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.201299 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.201330 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.201922 4867 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.66:6443: connect: connection refused" node="crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.272759 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.281265 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.302836 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.303401 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ead240e974956b450c9703a9e9b268d154ccce43af2727e4928f90a6452d7731 WatchSource:0}: Error finding container ead240e974956b450c9703a9e9b268d154ccce43af2727e4928f90a6452d7731: Status 404 returned error can't find the container with id ead240e974956b450c9703a9e9b268d154ccce43af2727e4928f90a6452d7731 Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.313028 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.313281 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-b132e74f6db54e2663aaf02b441bdf55251628a1601f707f6c49c23ea6f11c7b WatchSource:0}: Error finding container b132e74f6db54e2663aaf02b441bdf55251628a1601f707f6c49c23ea6f11c7b: Status 404 returned error can't find the container with id b132e74f6db54e2663aaf02b441bdf55251628a1601f707f6c49c23ea6f11c7b Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.317768 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.377509 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="800ms" Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.396349 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9e92616cc0b97583b014faa63cce9da01a12c9fb1cbb68212e0b1ad07e701dd1 WatchSource:0}: Error finding container 9e92616cc0b97583b014faa63cce9da01a12c9fb1cbb68212e0b1ad07e701dd1: Status 404 returned error can't find the container with id 9e92616cc0b97583b014faa63cce9da01a12c9fb1cbb68212e0b1ad07e701dd1 Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.399297 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-80d3d6b6fb37f646510755961abc8bf39cd771a2a96cf7e5cfb9c1ef6b0f784f WatchSource:0}: Error finding container 80d3d6b6fb37f646510755961abc8bf39cd771a2a96cf7e5cfb9c1ef6b0f784f: Status 404 returned error can't find the container with id 80d3d6b6fb37f646510755961abc8bf39cd771a2a96cf7e5cfb9c1ef6b0f784f Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.602339 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.604004 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.604060 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.604074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.604118 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.604822 4867 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.66:6443: connect: connection refused" node="crc" Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.697660 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.697759 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.761947 4867 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.762849 4867 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 02:24:42.363871588 +0000 UTC Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.762926 4867 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 547h36m8.600948634s for next certificate rotation Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.836555 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.836676 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.842478 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.842635 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e8c812323490eb761fe4a563939a3d5de0ccf9c72fe0b517a23b8569cca333b8"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.844950 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8" exitCode=0 Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.845036 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.845080 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"80d3d6b6fb37f646510755961abc8bf39cd771a2a96cf7e5cfb9c1ef6b0f784f"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.845216 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849774 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849788 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849816 4867 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635" exitCode=0 Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849891 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.849945 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9e92616cc0b97583b014faa63cce9da01a12c9fb1cbb68212e0b1ad07e701dd1"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.850108 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.850904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.850941 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.850957 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.851912 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852547 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852576 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852803 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4bfc4f87be3c2757f1ffc85e2ce404ae26f8dd491cc50a2e66b22168cdd6bb67" exitCode=0 Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852865 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4bfc4f87be3c2757f1ffc85e2ce404ae26f8dd491cc50a2e66b22168cdd6bb67"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.852967 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b132e74f6db54e2663aaf02b441bdf55251628a1601f707f6c49c23ea6f11c7b"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.853100 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.854161 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.854193 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.854206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.855917 4867 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf" exitCode=0 Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.855969 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.856011 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ead240e974956b450c9703a9e9b268d154ccce43af2727e4928f90a6452d7731"} Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.856132 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.860328 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.860365 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:33 crc kubenswrapper[4867]: I1212 06:48:33.860377 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:33 crc kubenswrapper[4867]: W1212 06:48:33.885840 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:33 crc kubenswrapper[4867]: E1212 06:48:33.885961 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:34 crc kubenswrapper[4867]: W1212 06:48:34.154681 4867 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:34 crc kubenswrapper[4867]: E1212 06:48:34.154778 4867 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:34 crc kubenswrapper[4867]: E1212 06:48:34.180856 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="1.6s" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.405152 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.406514 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.406566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.406579 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.406609 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:34 crc kubenswrapper[4867]: E1212 06:48:34.407167 4867 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.66:6443: connect: connection refused" node="crc" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.688101 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 12 06:48:34 crc kubenswrapper[4867]: E1212 06:48:34.689324 4867 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.66:6443: connect: connection refused" logger="UnhandledError" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.762101 4867 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.66:6443: connect: connection refused Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.861834 4867 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f" exitCode=0 Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.861893 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.862031 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.863208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.863261 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.863275 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.866921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3df6cdaeced3b9fe00c1949dda9abc34e0e948e24f3d8f88ddbd0b090c9d74cf"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.867014 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.867879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.867907 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.867919 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.869569 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.869599 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.869618 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.869704 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.870638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.870713 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.870727 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.871660 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.871690 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.871705 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.871715 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.872454 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.872499 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.872528 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874784 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874840 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874853 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874864 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874875 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817"} Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.874975 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.875663 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.875694 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:34 crc kubenswrapper[4867]: I1212 06:48:34.875709 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.878286 4867 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2" exitCode=0 Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.878348 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2"} Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.878458 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.878508 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879341 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879370 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879362 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:35 crc kubenswrapper[4867]: I1212 06:48:35.879434 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.007600 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.008735 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.008776 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.008786 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.008818 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887012 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d"} Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887090 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f"} Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887116 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64"} Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887134 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5"} Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887151 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8"} Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.887367 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.888526 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.888573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:36 crc kubenswrapper[4867]: I1212 06:48:36.888594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.146787 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.147408 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.148990 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.149041 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.149057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.238864 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.239186 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.241540 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.241606 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.241625 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.247451 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.889647 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.891439 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.891493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:37 crc kubenswrapper[4867]: I1212 06:48:37.891504 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.391687 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.391969 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.393294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.393324 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.393332 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.559440 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.804979 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.892752 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.893879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.893934 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:38 crc kubenswrapper[4867]: I1212 06:48:38.893946 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:39 crc kubenswrapper[4867]: I1212 06:48:39.247889 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:39 crc kubenswrapper[4867]: I1212 06:48:39.894585 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:39 crc kubenswrapper[4867]: I1212 06:48:39.895545 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:39 crc kubenswrapper[4867]: I1212 06:48:39.895574 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:39 crc kubenswrapper[4867]: I1212 06:48:39.895584 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.870436 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.870701 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.874851 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.874904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.874917 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.879438 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.897196 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.898357 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.898385 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:40 crc kubenswrapper[4867]: I1212 06:48:40.898397 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:41 crc kubenswrapper[4867]: I1212 06:48:41.864116 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 12 06:48:41 crc kubenswrapper[4867]: I1212 06:48:41.864470 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:41 crc kubenswrapper[4867]: I1212 06:48:41.866895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:41 crc kubenswrapper[4867]: I1212 06:48:41.866953 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:41 crc kubenswrapper[4867]: I1212 06:48:41.866979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:42 crc kubenswrapper[4867]: E1212 06:48:42.902699 4867 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.539754 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.539958 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.541108 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.541143 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.541155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.543708 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.905011 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.905755 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.905792 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:43 crc kubenswrapper[4867]: I1212 06:48:43.905804 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:44 crc kubenswrapper[4867]: I1212 06:48:44.780487 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 12 06:48:44 crc kubenswrapper[4867]: I1212 06:48:44.780743 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:44 crc kubenswrapper[4867]: I1212 06:48:44.781960 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:44 crc kubenswrapper[4867]: I1212 06:48:44.782016 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:44 crc kubenswrapper[4867]: I1212 06:48:44.782030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:45 crc kubenswrapper[4867]: I1212 06:48:45.276835 4867 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 12 06:48:45 crc kubenswrapper[4867]: I1212 06:48:45.276917 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 12 06:48:45 crc kubenswrapper[4867]: I1212 06:48:45.281917 4867 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 12 06:48:45 crc kubenswrapper[4867]: I1212 06:48:45.281991 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 12 06:48:46 crc kubenswrapper[4867]: I1212 06:48:46.540264 4867 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 12 06:48:46 crc kubenswrapper[4867]: I1212 06:48:46.540389 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.251214 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.252151 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.254132 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.254173 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.254184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.257469 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.917574 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.917628 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.918412 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.918479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:49 crc kubenswrapper[4867]: I1212 06:48:49.918497 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.265355 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.266937 4867 trace.go:236] Trace[1405015105]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (12-Dec-2025 06:48:36.521) (total time: 13745ms): Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[1405015105]: ---"Objects listed" error: 13745ms (06:48:50.266) Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[1405015105]: [13.745342434s] [13.745342434s] END Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.266963 4867 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.268310 4867 trace.go:236] Trace[2006742741]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (12-Dec-2025 06:48:36.909) (total time: 13359ms): Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[2006742741]: ---"Objects listed" error: 13358ms (06:48:50.268) Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[2006742741]: [13.359016092s] [13.359016092s] END Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.268326 4867 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.269474 4867 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.269546 4867 trace.go:236] Trace[630861029]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (12-Dec-2025 06:48:36.243) (total time: 14025ms): Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[630861029]: ---"Objects listed" error: 14025ms (06:48:50.269) Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[630861029]: [14.025621163s] [14.025621163s] END Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.269582 4867 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.269976 4867 trace.go:236] Trace[1590794755]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (12-Dec-2025 06:48:36.190) (total time: 14079ms): Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[1590794755]: ---"Objects listed" error: 14079ms (06:48:50.269) Dec 12 06:48:50 crc kubenswrapper[4867]: Trace[1590794755]: [14.079805504s] [14.079805504s] END Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.269996 4867 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.271279 4867 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.285724 4867 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.309302 4867 csr.go:261] certificate signing request csr-wk9wd is approved, waiting to be issued Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.316988 4867 csr.go:257] certificate signing request csr-wk9wd is issued Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.482730 4867 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.482803 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.482892 4867 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.482974 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.486494 4867 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60012->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.486584 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:60012->192.168.126.11:17697: read: connection reset by peer" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.757950 4867 apiserver.go:52] "Watching apiserver" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.762552 4867 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.762960 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dsjbr","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-njfkw"] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.763359 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.763530 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.763620 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.763898 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.763968 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.764054 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.764077 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.764089 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.764127 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.764096 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.764359 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.766759 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.766760 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.767118 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.767300 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.768463 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.768569 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.768733 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.768770 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.768817 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.769040 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.769318 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.769354 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.769768 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.770470 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.770551 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.771068 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.772218 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.772297 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.772337 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.772374 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.773169 4867 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.783618 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.784615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.785465 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.785558 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.785632 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.785753 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:51.285736954 +0000 UTC m=+18.857118223 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.797440 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.800874 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.811125 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.822486 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.831053 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.843033 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.854200 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.864072 4867 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.864626 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.873956 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874174 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874286 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874375 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874455 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874546 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874463 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874695 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874775 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874788 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874660 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874941 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.874968 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875004 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875030 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875050 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875067 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875088 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875107 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875123 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875142 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875165 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875181 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875200 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875216 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875248 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875265 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875287 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875316 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875336 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875355 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875394 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875500 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875520 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875537 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875555 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875573 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875593 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875615 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875635 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875658 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875677 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875699 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875722 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875744 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875768 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875793 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875812 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875863 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875884 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875905 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875924 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875946 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875969 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875994 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876015 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876034 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876088 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876108 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876128 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876150 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876172 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876192 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876213 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876250 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876311 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876331 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876355 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876382 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876403 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876425 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876445 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876468 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876488 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876507 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876526 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876545 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876567 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876591 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876669 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876690 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876708 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876726 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876750 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876769 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876788 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876805 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876825 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876842 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876860 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876877 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876894 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876912 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876933 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876950 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876968 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876988 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877005 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877021 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877040 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877059 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877076 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877092 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877113 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877129 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877146 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877162 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877178 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877195 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877212 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877244 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877261 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877279 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877298 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877315 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877333 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877352 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877369 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877390 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877446 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877465 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877484 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877502 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877520 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877540 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877557 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877574 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877595 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877613 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877630 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877646 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877667 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877685 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877731 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877846 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877865 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877892 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877908 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877926 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878023 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878043 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878060 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878076 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878093 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878110 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878126 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878143 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878161 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878178 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878219 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878251 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878268 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878284 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878301 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878317 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878335 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878352 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878369 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878405 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878421 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878438 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878454 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878474 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878494 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878511 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878529 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878548 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878565 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878604 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878623 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878639 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878658 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878674 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878691 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878708 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878727 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878745 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878766 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878804 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878822 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878840 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878873 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878891 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878909 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878928 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878946 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.878989 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879347 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879375 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879395 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879435 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879454 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879474 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879547 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879637 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/74c23654-e77b-459c-b6fe-aa39e7e0bed2-serviceca\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879661 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879696 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879754 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879775 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879815 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879841 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879860 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879879 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74c23654-e77b-459c-b6fe-aa39e7e0bed2-host\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879897 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx7q9\" (UniqueName: \"kubernetes.io/projected/74c23654-e77b-459c-b6fe-aa39e7e0bed2-kube-api-access-dx7q9\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879914 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c7878997-f37a-4633-ae33-391200ab7e05-hosts-file\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879948 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879971 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879992 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bsld\" (UniqueName: \"kubernetes.io/projected/c7878997-f37a-4633-ae33-391200ab7e05-kube-api-access-8bsld\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880063 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880075 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875025 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880889 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875306 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880905 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875464 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875580 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875635 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875673 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875698 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875811 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.875947 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876020 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876019 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876043 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876038 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876355 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876351 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876429 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876446 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876547 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876559 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876646 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876722 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.876801 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.881188 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877084 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877391 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877407 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877428 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877510 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.877792 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879151 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.879789 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880088 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880117 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880314 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880302 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880378 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880672 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.880811 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.881530 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.882306 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.883213 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.883745 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.883804 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.883886 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.883895 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884070 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884067 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884184 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884661 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884754 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884831 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.884890 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885038 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885380 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885467 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885498 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885539 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885549 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885588 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885591 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.885884 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886378 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886409 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886412 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886546 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886762 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.886796 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.887394 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.887474 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.887832 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.887878 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.888427 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.888873 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.888955 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.889332 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.889450 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.889516 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.890207 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.890304 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.890969 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.891348 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.891490 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.891610 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.891882 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.891987 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.892466 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.892536 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.893288 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.893869 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.893995 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.894237 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896292 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896274 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896351 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896373 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896398 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896468 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896538 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896687 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896695 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896832 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.896976 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897211 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897295 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897537 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897636 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897715 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897928 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897957 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.897994 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898334 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898396 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898614 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898660 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898712 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.898923 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.899208 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.899294 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.899527 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.899844 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.900043 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.900310 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.900364 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.901039 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.901098 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.901397 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.901452 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.901592 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:48:51.401566788 +0000 UTC m=+18.972948057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.901751 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.901840 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:51.401814505 +0000 UTC m=+18.973195774 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.902529 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.902676 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9dpkr"] Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.902833 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.902870 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:51.402861798 +0000 UTC m=+18.974243067 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.902954 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.903094 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.903182 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.903214 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-z8lss"] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.903352 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbsqz"] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.903752 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906187 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906211 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906384 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906445 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906546 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906591 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.906891 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.907652 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.908435 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.908427 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.909108 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.909531 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.905344 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.905355 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.911065 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.911306 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.911410 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.912182 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.912939 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-q8lbh"] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.913150 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.917173 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.917558 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.917831 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.919671 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.919931 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.920177 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.920257 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.920489 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.920558 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.920763 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:50 crc kubenswrapper[4867]: E1212 06:48:50.920928 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:51.420905104 +0000 UTC m=+18.992286383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.920261 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.922749 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.923036 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.923402 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.923609 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.924058 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.924208 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.924298 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.924362 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.925144 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.925318 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.925435 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.925611 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.925940 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.926445 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.926493 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.926622 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.926879 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.927557 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.927905 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928269 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928277 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928289 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928603 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928640 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928883 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.928949 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929022 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.930264 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929045 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929114 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929264 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929273 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.929284 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.933181 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.939791 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.940194 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.939042 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.940792 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.940812 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941173 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941276 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941502 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941685 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941757 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941677 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.941814 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.942377 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.942401 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.942375 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.944072 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.942520 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.945891 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.946192 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.946423 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947154 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947162 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947563 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2" exitCode=255 Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947613 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2"} Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947892 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.947978 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.957112 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.966734 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.969088 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.969597 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.975696 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980764 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980883 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980910 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980933 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980951 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-k8s-cni-cncf-io\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.980968 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-daemon-config\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981006 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9rjk\" (UniqueName: \"kubernetes.io/projected/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-kube-api-access-q9rjk\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981057 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981091 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-binary-copy\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-system-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981195 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-cnibin\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981241 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-os-release\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981295 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981320 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-hostroot\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981395 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-rootfs\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981417 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981451 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981474 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mhrh\" (UniqueName: \"kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981624 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bsld\" (UniqueName: \"kubernetes.io/projected/c7878997-f37a-4633-ae33-391200ab7e05-kube-api-access-8bsld\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981693 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981720 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-conf-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981746 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-etc-kubernetes\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981773 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981794 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981820 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981845 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981868 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981918 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981943 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-netns\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.981968 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-multus-certs\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982008 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-cnibin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982054 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982101 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982122 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982167 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982190 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982343 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74c23654-e77b-459c-b6fe-aa39e7e0bed2-host\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982453 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx7q9\" (UniqueName: \"kubernetes.io/projected/74c23654-e77b-459c-b6fe-aa39e7e0bed2-kube-api-access-dx7q9\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982501 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c7878997-f37a-4633-ae33-391200ab7e05-hosts-file\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982510 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74c23654-e77b-459c-b6fe-aa39e7e0bed2-host\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982531 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982562 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-os-release\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982588 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-proxy-tls\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982640 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-mcd-auth-proxy-config\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982662 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c7878997-f37a-4633-ae33-391200ab7e05-hosts-file\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982693 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-multus\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982763 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982795 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tmjr\" (UniqueName: \"kubernetes.io/projected/f1918c53-1a6d-4317-a320-2e301b5a9db7-kube-api-access-9tmjr\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982904 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.982955 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgnm\" (UniqueName: \"kubernetes.io/projected/a8287969-549f-480e-90bd-4478730313af-kube-api-access-wdgnm\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983089 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983169 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/74c23654-e77b-459c-b6fe-aa39e7e0bed2-serviceca\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983252 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983287 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-socket-dir-parent\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983411 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-kubelet\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983477 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-cni-binary-copy\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983497 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-bin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983530 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-system-cni-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983637 4867 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983651 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983662 4867 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983671 4867 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983684 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983694 4867 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983704 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983714 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983725 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983734 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983744 4867 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983754 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983763 4867 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983772 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983782 4867 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983791 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983799 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983810 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983819 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983828 4867 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983836 4867 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983845 4867 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983855 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983870 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983887 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983903 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983914 4867 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983924 4867 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983936 4867 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983947 4867 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983959 4867 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983970 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.983990 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/74c23654-e77b-459c-b6fe-aa39e7e0bed2-serviceca\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984093 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984110 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984138 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984212 4867 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984335 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984445 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984460 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984471 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984483 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984493 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984505 4867 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984516 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984527 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984538 4867 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984647 4867 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984679 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984691 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984703 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984715 4867 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984752 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984765 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984774 4867 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984783 4867 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984853 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984871 4867 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984881 4867 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984890 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984899 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984907 4867 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984918 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984931 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984971 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.984987 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985000 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985110 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985123 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985136 4867 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985167 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985179 4867 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985192 4867 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985204 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985216 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985244 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985257 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985270 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985284 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985295 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985307 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985319 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985330 4867 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985343 4867 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985356 4867 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985370 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985381 4867 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985394 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985405 4867 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985417 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985430 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985443 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985455 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985467 4867 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985479 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985492 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985504 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985516 4867 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985527 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985536 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985603 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985613 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985621 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985630 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985643 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985651 4867 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985660 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985669 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985677 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985686 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985695 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985705 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985714 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985723 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985731 4867 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985739 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985748 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985757 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985767 4867 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985778 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985790 4867 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985801 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985811 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985822 4867 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985832 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985844 4867 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985856 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985865 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985874 4867 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985883 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985891 4867 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985900 4867 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985909 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985924 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985932 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985940 4867 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985949 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985957 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985965 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985975 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985983 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985991 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.985999 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986007 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986015 4867 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986024 4867 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986032 4867 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986040 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986050 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986059 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986068 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986077 4867 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986085 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986094 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986103 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986112 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986121 4867 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986129 4867 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986137 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986146 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986154 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986163 4867 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986171 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986180 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986188 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986202 4867 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986210 4867 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986219 4867 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986242 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986250 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986258 4867 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986267 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986275 4867 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986284 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986292 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986301 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986311 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986319 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986327 4867 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986336 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986344 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986352 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986360 4867 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986369 4867 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986377 4867 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986385 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986393 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986402 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986411 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986419 4867 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986427 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986434 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.986443 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.990458 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.997383 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bsld\" (UniqueName: \"kubernetes.io/projected/c7878997-f37a-4633-ae33-391200ab7e05-kube-api-access-8bsld\") pod \"node-resolver-njfkw\" (UID: \"c7878997-f37a-4633-ae33-391200ab7e05\") " pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:50 crc kubenswrapper[4867]: I1212 06:48:50.997860 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx7q9\" (UniqueName: \"kubernetes.io/projected/74c23654-e77b-459c-b6fe-aa39e7e0bed2-kube-api-access-dx7q9\") pod \"node-ca-dsjbr\" (UID: \"74c23654-e77b-459c-b6fe-aa39e7e0bed2\") " pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.000728 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.009828 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.023496 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.032764 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.033591 4867 scope.go:117] "RemoveContainer" containerID="1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.035509 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.046594 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.054394 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.064734 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.076823 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.076916 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088152 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088429 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088684 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088756 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-os-release\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-multus\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088833 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-proxy-tls\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088865 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-mcd-auth-proxy-config\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088889 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tmjr\" (UniqueName: \"kubernetes.io/projected/f1918c53-1a6d-4317-a320-2e301b5a9db7-kube-api-access-9tmjr\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088931 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088950 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgnm\" (UniqueName: \"kubernetes.io/projected/a8287969-549f-480e-90bd-4478730313af-kube-api-access-wdgnm\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.088992 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089164 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089194 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-socket-dir-parent\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089218 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-kubelet\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089262 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-system-cni-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-cni-binary-copy\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-bin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089496 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089564 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-k8s-cni-cncf-io\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089631 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-daemon-config\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089651 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9rjk\" (UniqueName: \"kubernetes.io/projected/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-kube-api-access-q9rjk\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089685 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089706 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-binary-copy\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089725 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-system-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089771 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-cnibin\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089792 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-os-release\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089812 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089845 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-hostroot\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089860 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-rootfs\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089880 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.089955 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090012 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mhrh\" (UniqueName: \"kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090053 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-conf-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090073 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-etc-kubernetes\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090094 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090134 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090200 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090241 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090286 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090308 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090331 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-netns\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090346 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-multus-certs\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-cnibin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090413 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090433 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090488 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090846 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.090975 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-os-release\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.091010 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-multus\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.091941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-cnibin\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.092009 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-cni-bin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.092055 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.092615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-mcd-auth-proxy-config\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.092598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.092752 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-os-release\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093066 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-k8s-cni-cncf-io\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093145 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093099 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093187 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-cni-binary-copy\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093298 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-system-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093363 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-netns\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093425 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093460 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093494 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093528 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093745 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093851 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-run-multus-certs\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093896 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093929 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-cnibin\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.093850 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094087 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-cni-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094098 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-host-var-lib-kubelet\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094178 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-socket-dir-parent\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094245 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a8287969-549f-480e-90bd-4478730313af-system-cni-dir\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094329 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-hostroot\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094341 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094388 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094433 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-etc-kubernetes\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094447 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094456 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094404 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-rootfs\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094406 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.094495 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-conf-dir\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.095121 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-proxy-tls\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.095916 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/f1918c53-1a6d-4317-a320-2e301b5a9db7-multus-daemon-config\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.096131 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.096522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: W1212 06:48:51.097618 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-8125084ed36d326dfb1a08cbc9fa0b1cd7d26761ba271bcb99e2e94a69a9c38a WatchSource:0}: Error finding container 8125084ed36d326dfb1a08cbc9fa0b1cd7d26761ba271bcb99e2e94a69a9c38a: Status 404 returned error can't find the container with id 8125084ed36d326dfb1a08cbc9fa0b1cd7d26761ba271bcb99e2e94a69a9c38a Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.098442 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-njfkw" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.099434 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a8287969-549f-480e-90bd-4478730313af-cni-binary-copy\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.105427 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.110534 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tmjr\" (UniqueName: \"kubernetes.io/projected/f1918c53-1a6d-4317-a320-2e301b5a9db7-kube-api-access-9tmjr\") pod \"multus-z8lss\" (UID: \"f1918c53-1a6d-4317-a320-2e301b5a9db7\") " pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.112390 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.114491 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9rjk\" (UniqueName: \"kubernetes.io/projected/f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae-kube-api-access-q9rjk\") pod \"machine-config-daemon-q8lbh\" (UID: \"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\") " pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.115015 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgnm\" (UniqueName: \"kubernetes.io/projected/a8287969-549f-480e-90bd-4478730313af-kube-api-access-wdgnm\") pod \"multus-additional-cni-plugins-9dpkr\" (UID: \"a8287969-549f-480e-90bd-4478730313af\") " pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.117282 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mhrh\" (UniqueName: \"kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh\") pod \"ovnkube-node-zbsqz\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.119015 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.125423 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dsjbr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.136615 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.150265 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.162711 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.171282 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.181507 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.198361 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.210023 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.261196 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.264196 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z8lss" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.270335 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.276313 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.293816 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.294056 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.294083 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.294100 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.294173 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:52.29414679 +0000 UTC m=+19.865528059 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.319343 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-12 06:43:50 +0000 UTC, rotation deadline is 2026-09-12 23:19:59.842602621 +0000 UTC Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.319417 4867 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6592h31m8.523188073s for next certificate rotation Dec 12 06:48:51 crc kubenswrapper[4867]: W1212 06:48:51.333348 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8287969_549f_480e_90bd_4478730313af.slice/crio-c87d1e7dba55f8c19cdded159cec2209ed73f6cf21f073291eb6d01be34fd42b WatchSource:0}: Error finding container c87d1e7dba55f8c19cdded159cec2209ed73f6cf21f073291eb6d01be34fd42b: Status 404 returned error can't find the container with id c87d1e7dba55f8c19cdded159cec2209ed73f6cf21f073291eb6d01be34fd42b Dec 12 06:48:51 crc kubenswrapper[4867]: W1212 06:48:51.348186 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf99b86bf_5b7d_4b8a_ab9b_df4cd408ccae.slice/crio-87c22faf696174b091e9aabb57064a7601ad4997e407fac804fdff0d8a14236a WatchSource:0}: Error finding container 87c22faf696174b091e9aabb57064a7601ad4997e407fac804fdff0d8a14236a: Status 404 returned error can't find the container with id 87c22faf696174b091e9aabb57064a7601ad4997e407fac804fdff0d8a14236a Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.496064 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.496190 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.496219 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.496290 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496391 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496442 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:52.49642932 +0000 UTC m=+20.067810579 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496733 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496755 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496767 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496792 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496815 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:52.496797999 +0000 UTC m=+20.068179268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496834 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:52.49682608 +0000 UTC m=+20.068207349 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:51 crc kubenswrapper[4867]: E1212 06:48:51.496896 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:48:52.496891021 +0000 UTC m=+20.068272290 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.952357 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76" exitCode=0 Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.952449 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.952713 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerStarted","Data":"c87d1e7dba55f8c19cdded159cec2209ed73f6cf21f073291eb6d01be34fd42b"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.954417 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-njfkw" event={"ID":"c7878997-f37a-4633-ae33-391200ab7e05","Type":"ContainerStarted","Data":"72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.954492 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-njfkw" event={"ID":"c7878997-f37a-4633-ae33-391200ab7e05","Type":"ContainerStarted","Data":"f2d31c6ed9ae153736fdb1512bc7566e49609d5f9c0e759e0db81e9905a7bb73"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.955636 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.955664 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8125084ed36d326dfb1a08cbc9fa0b1cd7d26761ba271bcb99e2e94a69a9c38a"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.958832 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" exitCode=0 Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.958906 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.958929 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"b1632ebb28e10ceae66726db235e73630a37f4db4503e12ffbdb5a55ca91a407"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.961133 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dsjbr" event={"ID":"74c23654-e77b-459c-b6fe-aa39e7e0bed2","Type":"ContainerStarted","Data":"23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.961171 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dsjbr" event={"ID":"74c23654-e77b-459c-b6fe-aa39e7e0bed2","Type":"ContainerStarted","Data":"be53224b53b0f2ace00c9d08a42172e4c7a7870a2f67aa9ab746d6b8f916ffc7"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.964244 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.966490 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.966766 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.968283 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7353a7fc6a864636be8221ad7528405bc80f7fdef3872b74a39ff89d9afd76c8"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.969836 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.969894 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.969908 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c18961a9ec32955fd3e526bc542db6763eba00cb72e499d66908f7cbd7cc3461"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.970955 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerStarted","Data":"e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.971004 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerStarted","Data":"318087d8065141430ca9fbc91e6606fb3e4b2e90af8268e201d253b034d5b3a4"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.972685 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.972710 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.972722 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"87c22faf696174b091e9aabb57064a7601ad4997e407fac804fdff0d8a14236a"} Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.984731 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:51Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:51 crc kubenswrapper[4867]: I1212 06:48:51.999005 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:51Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.017054 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.030248 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.045867 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.059642 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.072938 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.085508 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.097891 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.112768 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.125547 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.137354 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.152007 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.168826 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.189283 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.200612 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.222659 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.239454 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.252455 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.266971 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.282274 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.295060 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.304054 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.304216 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.304256 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.304266 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.304328 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:54.304298693 +0000 UTC m=+21.875679962 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.310137 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.333972 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.354725 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.369155 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.507652 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.507796 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:48:54.507772551 +0000 UTC m=+22.079153820 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.508080 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.508119 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.508147 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508290 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508304 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508315 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508353 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:54.508343213 +0000 UTC m=+22.079724482 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508671 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508701 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:54.508692041 +0000 UTC m=+22.080073320 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508733 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.508765 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:54.508757152 +0000 UTC m=+22.080138431 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.677860 4867 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.837841 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.837897 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.837919 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.837997 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.838054 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:48:52 crc kubenswrapper[4867]: E1212 06:48:52.838171 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.841656 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.842781 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.843968 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.844781 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.845996 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.846673 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.847718 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.848942 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.849856 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.850260 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.851103 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.852143 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.853182 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.854195 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.854784 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.857122 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.857815 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.858498 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.859477 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.860102 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.860953 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.861958 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.862634 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.863616 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.863979 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.864453 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.864910 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.866146 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.867436 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.868162 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.868791 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.869816 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.870455 4867 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.870581 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.873183 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.873881 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.874469 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.876360 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.876871 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.877589 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.878176 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.879478 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.880134 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.880967 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.881689 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.882691 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.883314 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.884103 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.884640 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.885571 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.886938 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.887669 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.888589 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.889086 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.889135 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.889773 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.890860 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.891372 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.901888 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.915862 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.930255 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.944125 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.956794 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.968887 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.979056 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.979099 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.979110 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.979119 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.979129 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.980680 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112" exitCode=0 Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.981290 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112"} Dec 12 06:48:52 crc kubenswrapper[4867]: I1212 06:48:52.981967 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.000683 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.018388 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.030866 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.042274 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.055367 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.067324 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.081800 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.093111 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.110836 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.128319 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.143144 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.159085 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.171762 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.188111 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.199798 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.471326 4867 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.473327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.473370 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.473381 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.473541 4867 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.479755 4867 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.479978 4867 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.480766 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.480801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.480816 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.480831 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.480844 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.498855 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.502970 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.503003 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.503012 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.503026 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.503036 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.517569 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.522501 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.522565 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.522577 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.522598 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.522611 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.536853 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.540484 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.540537 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.540549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.540566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.540577 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.543138 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.546883 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.551896 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.557746 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.559790 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.561834 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.561870 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.561882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.561900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.561915 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.574368 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: E1212 06:48:53.574509 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.574599 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.576438 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.576472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.576482 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.576498 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.576509 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.597668 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.611522 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.624810 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.636651 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.649169 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.663535 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.676523 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.678364 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.678397 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.678406 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.678423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.678435 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.687785 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.706357 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.728737 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.741929 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.755130 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.780572 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.780608 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.780617 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.780632 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.780641 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.788837 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.827728 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.869505 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.882469 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.882510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.882520 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.882537 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.882547 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.908962 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.948027 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985344 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985412 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985422 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985440 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985453 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:53Z","lastTransitionTime":"2025-12-12T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985793 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6" exitCode=0 Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.985833 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.989921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 06:48:53 crc kubenswrapper[4867]: I1212 06:48:53.990662 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.030544 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.076347 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.087488 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.087529 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.087542 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.087557 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.087566 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.110493 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.149792 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.187640 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.189417 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.189457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.189471 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.189489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.189501 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.227981 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.272377 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.292072 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.292109 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.292120 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.292138 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.292148 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.310815 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.324708 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.324959 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.325001 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.325019 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.325092 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:58.325068371 +0000 UTC m=+25.896449800 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.352356 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.389410 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.394689 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.394720 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.394727 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.394742 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.394751 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.439285 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.478315 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.497036 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.497074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.497084 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.497099 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.497110 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.508712 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.526593 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.526718 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:48:58.526696737 +0000 UTC m=+26.098078146 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.526804 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.526871 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.526907 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.526946 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.526992 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:58.526979513 +0000 UTC m=+26.098360782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527016 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527029 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527039 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527038 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527069 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:58.527062745 +0000 UTC m=+26.098444254 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.527139 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:48:58.527117166 +0000 UTC m=+26.098498435 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.547537 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.586173 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.599575 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.599613 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.599624 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.599642 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.599651 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.627999 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.669490 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.701840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.701875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.701885 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.701900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.701914 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.713874 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.751041 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.788064 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.804745 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.804812 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.804822 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.804838 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.804850 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.819424 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.834256 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.837346 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.837432 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.837377 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.837471 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.837565 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.837671 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:48:54 crc kubenswrapper[4867]: E1212 06:48:54.837836 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.856853 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.907576 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.907631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.907650 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.907678 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.907699 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:54Z","lastTransitionTime":"2025-12-12T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.909598 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.930371 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.970885 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:54Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.997817 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df" exitCode=0 Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.997974 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df"} Dec 12 06:48:54 crc kubenswrapper[4867]: I1212 06:48:54.999768 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.010310 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.010353 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.010365 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.010382 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.010397 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.024796 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: E1212 06:48:55.052181 4867 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.074537 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.111340 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.112823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.112853 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.112866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.112883 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.112894 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.147219 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.189120 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.215552 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.215590 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.215603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.215623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.215638 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.231104 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.271131 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.310396 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.317984 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.318021 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.318030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.318043 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.318052 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.350132 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.389115 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.420515 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.420553 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.420566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.420581 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.420593 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.431662 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.475895 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.511011 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.522899 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.522944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.522954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.522973 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.522983 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.552559 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.602532 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.626045 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.626093 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.626116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.626135 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.626148 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.631538 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.672388 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.713443 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.730472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.730534 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.730555 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.730584 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.730603 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.754708 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.800926 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.828263 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.833739 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.833787 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.833796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.833810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.833836 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.870405 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.915370 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.938203 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.938305 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.938329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.938365 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.938392 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:55Z","lastTransitionTime":"2025-12-12T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.951820 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:55 crc kubenswrapper[4867]: I1212 06:48:55.998492 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.008047 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.010275 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b" exitCode=0 Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.011364 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.031193 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.046267 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.046298 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.046328 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.046343 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.046353 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.071898 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.114091 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.153275 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.153329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.153341 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.153359 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.153371 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.165978 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.187085 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.236833 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.256521 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.256574 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.256586 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.256605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.256617 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.266602 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.309696 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.349916 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.358772 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.358855 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.358879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.358907 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.358925 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.390800 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.430886 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.461535 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.461586 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.461601 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.461620 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.461632 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.470908 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.512283 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.552665 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.564960 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.565019 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.565030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.565051 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.565066 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.598746 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.626987 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:56Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.667674 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.667708 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.667719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.667738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.667751 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.771765 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.771826 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.771845 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.771878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.771902 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.837918 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.838027 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:56 crc kubenswrapper[4867]: E1212 06:48:56.838090 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.838117 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:56 crc kubenswrapper[4867]: E1212 06:48:56.838309 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:48:56 crc kubenswrapper[4867]: E1212 06:48:56.838483 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.874350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.874399 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.874410 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.874430 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.874442 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.977781 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.977841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.977851 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.977872 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:56 crc kubenswrapper[4867]: I1212 06:48:56.977882 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:56Z","lastTransitionTime":"2025-12-12T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.024823 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8287969-549f-480e-90bd-4478730313af" containerID="8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a" exitCode=0 Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.024894 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerDied","Data":"8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.059794 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.077005 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.081443 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.081521 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.081555 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.081592 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.081614 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.094731 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.122561 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.135849 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.150519 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.163571 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.177569 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.185702 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.185733 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.185742 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.185756 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.185765 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.193808 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.207911 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.226790 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.244374 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.258771 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.279736 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291315 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:57Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291577 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291621 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291658 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.291669 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.394586 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.394935 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.394947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.394969 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.394983 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.498454 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.498560 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.498579 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.498605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.498623 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.601535 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.601575 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.601585 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.601602 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.601612 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.704556 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.704598 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.704608 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.704623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.704633 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.808493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.808553 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.808566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.808592 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.808608 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.911565 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.911630 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.911646 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.911671 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:57 crc kubenswrapper[4867]: I1212 06:48:57.911685 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:57Z","lastTransitionTime":"2025-12-12T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.015097 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.015146 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.015156 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.015172 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.015184 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.032519 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.033008 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.040421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" event={"ID":"a8287969-549f-480e-90bd-4478730313af","Type":"ContainerStarted","Data":"e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.047598 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.058861 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.067743 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.083316 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.098832 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.117364 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.118548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.118587 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.118597 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.118614 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.118625 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.128793 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.134804 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.154017 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.163944 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.175601 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.187646 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.198635 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.208923 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.219752 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.220774 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.220802 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.220814 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.220831 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.220843 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.231016 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.242048 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.257544 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.265995 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.275067 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.285378 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.295541 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.305411 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.316993 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.323485 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.323526 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.323536 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.323550 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.323559 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.326213 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.338053 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.349601 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.359604 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.369553 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.392594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.392800 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.392832 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.392842 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.392899 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.392884367 +0000 UTC m=+33.964265636 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.396308 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.426181 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.426215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.426234 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.426248 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.426260 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.432975 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:58Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.529166 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.529400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.529409 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.529426 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.529438 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.594507 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.594620 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.594654 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594762 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.594730367 +0000 UTC m=+34.166111636 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594765 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594835 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.594824859 +0000 UTC m=+34.166206248 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594861 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594881 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594894 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594948 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.594725 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.594952 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.594936511 +0000 UTC m=+34.166317850 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.595040 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.595018333 +0000 UTC m=+34.166399602 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.632149 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.632216 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.632286 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.632317 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.632340 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.735180 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.735294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.735323 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.735353 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.735378 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837123 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.837285 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837336 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837361 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.837414 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:48:58 crc kubenswrapper[4867]: E1212 06:48:58.837476 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837818 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837872 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837890 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837912 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.837933 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.941009 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.941051 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.941059 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.941074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:58 crc kubenswrapper[4867]: I1212 06:48:58.941085 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:58Z","lastTransitionTime":"2025-12-12T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.042971 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.043010 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.043022 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.043039 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.043050 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.044044 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.044506 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.066033 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.079872 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.094572 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.109565 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.121171 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.136705 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.145754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.145787 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.145799 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.145817 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.145841 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.155939 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.174555 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.189324 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.207347 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.219902 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.231354 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.244390 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.247938 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.248014 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.248029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.248048 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.248059 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.257535 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.269947 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.282460 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:48:59Z is after 2025-08-24T17:21:41Z" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.351029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.351108 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.351122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.351145 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.351161 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.454129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.454177 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.454189 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.454209 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.454225 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.556098 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.556138 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.556151 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.556170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.556183 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.658262 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.658299 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.658308 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.658342 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.658354 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.760271 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.760311 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.760323 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.760338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.760349 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.842255 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.862267 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.862308 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.862317 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.862333 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.862342 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.964878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.964964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.964978 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.964999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:48:59 crc kubenswrapper[4867]: I1212 06:48:59.965011 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:48:59Z","lastTransitionTime":"2025-12-12T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.067430 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.067463 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.067477 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.067493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.067521 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.169223 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.169467 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.169477 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.169490 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.169498 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.271272 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.271317 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.271327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.271341 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.271350 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.373800 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.373836 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.373847 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.373864 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.373876 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.475947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.475999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.476009 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.476023 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.476033 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.578180 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.578239 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.578252 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.578268 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.578279 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.680944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.680986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.680999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.681024 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.681035 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.783703 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.783741 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.783749 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.783763 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.783772 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.851793 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:00 crc kubenswrapper[4867]: E1212 06:49:00.851948 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.852379 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.852418 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:00 crc kubenswrapper[4867]: E1212 06:49:00.852505 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:00 crc kubenswrapper[4867]: E1212 06:49:00.852653 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.885800 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.886038 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.886122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.886204 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.886305 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.988882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.988922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.988931 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.988947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:00 crc kubenswrapper[4867]: I1212 06:49:00.988957 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:00Z","lastTransitionTime":"2025-12-12T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.050850 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/0.log" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.053297 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada" exitCode=1 Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.053414 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.054033 4867 scope.go:117] "RemoveContainer" containerID="4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.071448 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.084521 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.091216 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.091269 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.091279 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.091293 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.091309 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.095534 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.109658 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.128340 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:00Z\\\",\\\"message\\\":\\\"06:49:00.048148 6178 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1212 06:49:00.049425 6178 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:00.049446 6178 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:00.049468 6178 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:00.049475 6178 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:00.049575 6178 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:00.049614 6178 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:00.049621 6178 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:00.049631 6178 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1212 06:49:00.049644 6178 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1212 06:49:00.049649 6178 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:00.049660 6178 factory.go:656] Stopping watch factory\\\\nI1212 06:49:00.049671 6178 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:00.049674 6178 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1212 06:49:00.049683 6178 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:00.049681 6178 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.140411 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.153209 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.163968 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.176588 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.189367 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.193474 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.193513 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.193525 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.193544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.193556 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.211968 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.226059 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.237601 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.247871 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.258503 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:01Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.295439 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.295474 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.295486 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.295501 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.295510 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.397687 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.397716 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.397724 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.397738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.397748 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.499754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.499792 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.499802 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.499818 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.499828 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.602388 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.602603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.602687 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.602809 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.602893 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.705393 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.705443 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.705453 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.705472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.705486 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.807895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.807941 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.807950 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.807966 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.807978 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.910705 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.910754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.910767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.910785 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:01 crc kubenswrapper[4867]: I1212 06:49:01.910797 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:01Z","lastTransitionTime":"2025-12-12T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.012712 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.012755 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.012766 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.012782 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.012796 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.057949 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/0.log" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.061819 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.062184 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.079567 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.094841 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.113918 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:00Z\\\",\\\"message\\\":\\\"06:49:00.048148 6178 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1212 06:49:00.049425 6178 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:00.049446 6178 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:00.049468 6178 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:00.049475 6178 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:00.049575 6178 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:00.049614 6178 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:00.049621 6178 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:00.049631 6178 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1212 06:49:00.049644 6178 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1212 06:49:00.049649 6178 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:00.049660 6178 factory.go:656] Stopping watch factory\\\\nI1212 06:49:00.049671 6178 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:00.049674 6178 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1212 06:49:00.049683 6178 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:00.049681 6178 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.114695 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.114735 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.114744 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.114761 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.114772 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.127578 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.149374 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.158654 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.168301 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.184782 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.197337 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.209342 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.217475 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.217514 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.217523 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.217543 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.217552 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.224670 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.238672 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.260119 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.277060 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.288558 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.320727 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.320765 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.320774 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.320789 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.320800 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.423925 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.423986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.424002 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.424028 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.424047 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.526541 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.526575 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.526587 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.526601 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.526610 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.628985 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.629077 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.629093 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.629123 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.629142 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.731045 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.731098 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.731115 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.731138 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.731156 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.833638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.833688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.833699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.833713 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.833723 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.838015 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.838088 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.838067 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:02 crc kubenswrapper[4867]: E1212 06:49:02.838295 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:02 crc kubenswrapper[4867]: E1212 06:49:02.838318 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:02 crc kubenswrapper[4867]: E1212 06:49:02.838365 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.851345 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.862716 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.872205 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.883313 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.895619 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.911790 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.922019 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.933415 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.936203 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.936331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.936390 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.936462 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.936521 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:02Z","lastTransitionTime":"2025-12-12T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.944385 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.963017 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:00Z\\\",\\\"message\\\":\\\"06:49:00.048148 6178 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1212 06:49:00.049425 6178 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:00.049446 6178 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:00.049468 6178 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:00.049475 6178 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:00.049575 6178 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:00.049614 6178 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:00.049621 6178 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:00.049631 6178 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1212 06:49:00.049644 6178 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1212 06:49:00.049649 6178 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:00.049660 6178 factory.go:656] Stopping watch factory\\\\nI1212 06:49:00.049671 6178 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:00.049674 6178 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1212 06:49:00.049683 6178 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:00.049681 6178 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.973918 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:02 crc kubenswrapper[4867]: I1212 06:49:02.994808 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:02Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.005733 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.016484 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.031057 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.039109 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.039146 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.039160 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.039177 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.039187 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:03Z","lastTransitionTime":"2025-12-12T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.066354 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/1.log" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.066909 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/0.log" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.069052 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b" exitCode=1 Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.069093 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b"} Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.069144 4867 scope.go:117] "RemoveContainer" containerID="4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.069740 4867 scope.go:117] "RemoveContainer" containerID="b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b" Dec 12 06:49:03 crc kubenswrapper[4867]: E1212 06:49:03.069892 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.084373 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.095940 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.112523 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f640de32d03f01e0fdbdff30a2a985e85244952ef07dd94635853b87c330ada\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:00Z\\\",\\\"message\\\":\\\"06:49:00.048148 6178 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1212 06:49:00.049425 6178 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:00.049446 6178 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:00.049468 6178 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:00.049475 6178 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:00.049575 6178 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:00.049614 6178 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:00.049621 6178 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:00.049631 6178 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1212 06:49:00.049644 6178 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1212 06:49:00.049649 6178 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:00.049660 6178 factory.go:656] Stopping watch factory\\\\nI1212 06:49:00.049671 6178 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:00.049674 6178 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1212 06:49:00.049683 6178 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:00.049681 6178 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.122493 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.140990 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.141030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.141038 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.141051 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.141060 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:03Z","lastTransitionTime":"2025-12-12T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.145029 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:03Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.243461 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.243509 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.243525 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.243541 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.243553 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:03Z","lastTransitionTime":"2025-12-12T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.346095 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.346337 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.346428 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.346533 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:03 crc kubenswrapper[4867]: I1212 06:49:03.346596 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:03Z","lastTransitionTime":"2025-12-12T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.096150 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.096458 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.096768 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz"] Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.097478 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.100539 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.100569 4867 scope.go:117] "RemoveContainer" containerID="b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.100902 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.101188 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.101217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.101271 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.101290 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.101673 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.102010 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.103692 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.103751 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.103766 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.103783 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.103793 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.108485 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.117837 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.122756 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.122790 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.122799 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.122815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.122827 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.126246 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.141793 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.142759 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.146306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.146507 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.146531 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.146557 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.146573 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.155978 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.162078 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.167707 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.167788 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.167815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.167844 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.167869 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.170412 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.181019 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.183206 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.185951 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.186005 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.186046 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.186066 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.186078 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.195129 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.196922 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.197927 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.198315 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5hn9\" (UniqueName: \"kubernetes.io/projected/d604431e-f3e6-483a-abcc-92c8c5a5ab82-kube-api-access-f5hn9\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.198865 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.199024 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.199351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.203911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.203943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.203954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.203973 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.203987 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.208128 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.219952 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.230613 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.241929 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.254194 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.264322 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.278075 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.290587 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.301487 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.301589 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.301621 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.302376 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.302427 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.302446 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5hn9\" (UniqueName: \"kubernetes.io/projected/d604431e-f3e6-483a-abcc-92c8c5a5ab82-kube-api-access-f5hn9\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.302681 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307292 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d604431e-f3e6-483a-abcc-92c8c5a5ab82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307837 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307858 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307867 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.307914 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.314062 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.322942 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.333305 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5hn9\" (UniqueName: \"kubernetes.io/projected/d604431e-f3e6-483a-abcc-92c8c5a5ab82-kube-api-access-f5hn9\") pod \"ovnkube-control-plane-749d76644c-vrtlz\" (UID: \"d604431e-f3e6-483a-abcc-92c8c5a5ab82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.335685 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.351936 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.382278 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.396277 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.410408 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.410436 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.410445 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.410459 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.410468 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.415979 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.426170 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: W1212 06:49:04.437644 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd604431e_f3e6_483a_abcc_92c8c5a5ab82.slice/crio-ed2491c23a201d500368c773ecb7a4915e7c0da64d3d57b65323070260c04624 WatchSource:0}: Error finding container ed2491c23a201d500368c773ecb7a4915e7c0da64d3d57b65323070260c04624: Status 404 returned error can't find the container with id ed2491c23a201d500368c773ecb7a4915e7c0da64d3d57b65323070260c04624 Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.440571 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.456823 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.476089 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.512402 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.512444 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.512457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.512475 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.512488 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.615394 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.615457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.615470 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.615489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.615503 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.653862 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hj2lv"] Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.654457 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.654524 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.700530 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.706893 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.706965 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssjqs\" (UniqueName: \"kubernetes.io/projected/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-kube-api-access-ssjqs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.713375 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.717399 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.717431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.717440 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.717457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.717466 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.725958 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.738779 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.749944 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.763364 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.773073 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.789160 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.799025 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.808541 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.808580 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssjqs\" (UniqueName: \"kubernetes.io/projected/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-kube-api-access-ssjqs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.808701 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.808768 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:05.30874869 +0000 UTC m=+32.880129959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.813627 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.819520 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.819562 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.819573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.819591 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.819603 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.823954 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.824142 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssjqs\" (UniqueName: \"kubernetes.io/projected/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-kube-api-access-ssjqs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.833684 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.837426 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.837500 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.837598 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:04 crc kubenswrapper[4867]: E1212 06:49:04.837720 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.844664 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.854688 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.863950 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.871596 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.882576 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:04Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.922214 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.922270 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.922282 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.922303 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:04 crc kubenswrapper[4867]: I1212 06:49:04.922315 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:04Z","lastTransitionTime":"2025-12-12T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.025091 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.025421 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.025433 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.025451 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.025463 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.105118 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/1.log" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.108903 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" event={"ID":"d604431e-f3e6-483a-abcc-92c8c5a5ab82","Type":"ContainerStarted","Data":"3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.108938 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" event={"ID":"d604431e-f3e6-483a-abcc-92c8c5a5ab82","Type":"ContainerStarted","Data":"445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.108947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" event={"ID":"d604431e-f3e6-483a-abcc-92c8c5a5ab82","Type":"ContainerStarted","Data":"ed2491c23a201d500368c773ecb7a4915e7c0da64d3d57b65323070260c04624"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.125038 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.127298 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.127327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.127338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.127355 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.127366 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.141405 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.166820 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.180003 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.190162 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.212074 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.222059 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.229561 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.229584 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.229592 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.229605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.229616 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.232471 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.244687 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.254992 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.265719 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.276689 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.288374 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.298747 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.313564 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:05 crc kubenswrapper[4867]: E1212 06:49:05.313682 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:05 crc kubenswrapper[4867]: E1212 06:49:05.313731 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:06.313717519 +0000 UTC m=+33.885098788 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.316228 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.329099 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.332265 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.332296 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.332318 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.332338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.332352 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.336725 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:05Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.434878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.434914 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.434926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.434943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.434954 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.536582 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.536625 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.536638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.536655 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.536667 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.638530 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.638564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.638578 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.638623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.638632 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.740603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.740637 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.740646 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.740675 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.740684 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.837069 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:05 crc kubenswrapper[4867]: E1212 06:49:05.837308 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.843499 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.843555 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.843572 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.843596 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.843612 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.945431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.945466 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.945478 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.945495 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:05 crc kubenswrapper[4867]: I1212 06:49:05.945508 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:05Z","lastTransitionTime":"2025-12-12T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.049634 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.049699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.049723 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.049750 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.049771 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.152141 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.152184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.152213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.152242 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.152252 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.255143 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.255186 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.255198 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.255215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.255251 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.322946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.323142 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.323265 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:08.323245728 +0000 UTC m=+35.894626997 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.357118 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.357173 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.357189 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.357210 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.357255 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.424094 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.424376 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.424405 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.424426 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.424506 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:22.424482995 +0000 UTC m=+49.995864304 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.459230 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.459307 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.459321 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.459341 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.459354 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.562056 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.562101 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.562116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.562137 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.562187 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.626431 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.626508 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.626543 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.626577 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626648 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626685 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:49:22.626650984 +0000 UTC m=+50.198032253 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626714 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626735 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:22.626719765 +0000 UTC m=+50.198101154 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626745 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626750 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626845 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:22.626828708 +0000 UTC m=+50.198209977 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626757 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.626903 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:22.626892969 +0000 UTC m=+50.198274238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.664879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.664922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.664933 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.664951 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.664964 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.768019 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.768071 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.768083 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.768103 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.768113 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.837309 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.837401 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.837535 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.837548 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.837634 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:06 crc kubenswrapper[4867]: E1212 06:49:06.837714 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.870774 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.870806 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.870816 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.870836 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.870847 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.973428 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.973460 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.973470 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.973489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:06 crc kubenswrapper[4867]: I1212 06:49:06.973499 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:06Z","lastTransitionTime":"2025-12-12T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.075849 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.075887 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.075899 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.075915 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.075929 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.178128 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.178163 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.178171 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.178184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.178193 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.279887 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.279912 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.279921 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.279933 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.279942 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.381930 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.381969 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.381979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.381995 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.382006 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.483996 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.484295 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.484384 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.484504 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.484591 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.586598 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.586882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.586966 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.587047 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.587135 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.689954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.689999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.690011 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.690029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.690040 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.792710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.792747 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.792754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.792767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.792777 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.837830 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:07 crc kubenswrapper[4867]: E1212 06:49:07.837948 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.895274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.895318 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.895332 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.895350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.895363 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.997625 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.997658 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.997671 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.997689 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:07 crc kubenswrapper[4867]: I1212 06:49:07.997742 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:07Z","lastTransitionTime":"2025-12-12T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.100032 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.100070 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.100080 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.100096 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.100109 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.202902 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.203666 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.203817 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.203944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.204052 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.306376 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.306439 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.306457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.306479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.306496 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.343833 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:08 crc kubenswrapper[4867]: E1212 06:49:08.344042 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:08 crc kubenswrapper[4867]: E1212 06:49:08.344106 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:12.344084737 +0000 UTC m=+39.915466046 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.398594 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.409528 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.409596 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.409617 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.409684 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.409704 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.414754 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.427995 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.450156 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.463536 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.490060 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.503744 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.511847 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.511882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.511891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.511905 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.511915 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.521089 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.541816 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.554490 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.570113 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.582090 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.595265 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.613174 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.614080 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.614099 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.614107 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.614119 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.614128 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.622846 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.633993 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.644107 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.652521 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:08Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.716612 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.716856 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.716941 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.717029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.717112 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.819155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.819197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.819206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.819238 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.819249 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.837467 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:08 crc kubenswrapper[4867]: E1212 06:49:08.837617 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.837678 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.837713 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:08 crc kubenswrapper[4867]: E1212 06:49:08.837786 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:08 crc kubenswrapper[4867]: E1212 06:49:08.837856 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.921034 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.921062 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.921070 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.921082 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:08 crc kubenswrapper[4867]: I1212 06:49:08.921092 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:08Z","lastTransitionTime":"2025-12-12T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.023672 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.023710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.023720 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.023736 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.023745 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.126717 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.126799 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.126830 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.126860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.126884 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.229321 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.229362 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.229381 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.229400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.229411 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.331805 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.332117 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.332342 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.332513 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.332634 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.435040 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.435075 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.435086 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.435100 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.435110 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.538635 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.538688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.538699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.538719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.538731 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.641912 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.641959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.641972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.641992 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.642006 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.744665 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.744711 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.744725 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.744745 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.744758 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.837672 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:09 crc kubenswrapper[4867]: E1212 06:49:09.838117 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.848420 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.848464 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.848505 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.848531 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.848550 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.951989 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.952044 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.952064 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.952089 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:09 crc kubenswrapper[4867]: I1212 06:49:09.952109 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:09Z","lastTransitionTime":"2025-12-12T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.054653 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.054706 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.054724 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.054749 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.054768 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.157494 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.157588 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.157611 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.157640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.157659 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.261539 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.261609 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.261627 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.261653 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.261675 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.364875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.365360 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.365549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.365703 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.365880 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.468981 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.469093 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.469123 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.469159 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.469276 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.572020 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.572669 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.572807 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.572979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.573102 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.676476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.676592 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.676620 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.676657 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.676676 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.787359 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.787436 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.787459 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.787489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.787515 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.838154 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.838327 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:10 crc kubenswrapper[4867]: E1212 06:49:10.838401 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.838466 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:10 crc kubenswrapper[4867]: E1212 06:49:10.838586 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:10 crc kubenswrapper[4867]: E1212 06:49:10.838627 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.890448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.891148 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.891837 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.892393 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.892934 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.995969 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.996246 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.996423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.996597 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:10 crc kubenswrapper[4867]: I1212 06:49:10.996718 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:10Z","lastTransitionTime":"2025-12-12T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.099692 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.099732 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.099784 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.099812 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.099825 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.202453 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.202754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.203094 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.203468 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.203843 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.305871 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.305919 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.305932 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.305951 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.305963 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.408544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.408622 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.408640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.408677 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.408702 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.511810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.511861 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.511877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.511900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.511919 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.614909 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.615357 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.615589 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.616209 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.616832 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.719708 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.720050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.720153 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.720281 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.720396 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.823067 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.823107 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.823116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.823133 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.823145 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.837439 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:11 crc kubenswrapper[4867]: E1212 06:49:11.837567 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.925895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.925960 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.925975 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.925999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:11 crc kubenswrapper[4867]: I1212 06:49:11.926014 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:11Z","lastTransitionTime":"2025-12-12T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.029161 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.029194 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.029202 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.029220 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.029242 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.131511 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.131585 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.131603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.131631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.131650 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.234926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.234989 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.235003 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.235023 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.235039 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.339014 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.339056 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.339069 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.339089 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.339100 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.384281 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:12 crc kubenswrapper[4867]: E1212 06:49:12.384586 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:12 crc kubenswrapper[4867]: E1212 06:49:12.384749 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:20.384705859 +0000 UTC m=+47.956087158 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.442493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.442567 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.442591 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.442634 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.442658 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.546081 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.546159 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.546179 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.546206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.546259 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.649770 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.650160 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.650415 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.650599 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.650738 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.754408 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.754775 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.754937 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.755067 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.755188 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.838105 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.838168 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.838219 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:12 crc kubenswrapper[4867]: E1212 06:49:12.838893 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:12 crc kubenswrapper[4867]: E1212 06:49:12.838533 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:12 crc kubenswrapper[4867]: E1212 06:49:12.838994 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.857936 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.858322 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.858509 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.858536 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.858568 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.858592 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.877695 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.900902 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.923412 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.940597 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.959798 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.961882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.961929 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.961938 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.961976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.961986 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:12Z","lastTransitionTime":"2025-12-12T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:12 crc kubenswrapper[4867]: I1212 06:49:12.978279 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.000816 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:12Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.030250 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.044488 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.064175 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.065608 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.065651 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.065663 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.065683 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.065694 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.075019 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.086199 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.099073 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.112317 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.123734 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.137374 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:13Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.168713 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.168758 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.168768 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.168786 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.168798 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.271529 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.271762 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.271850 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.271959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.272033 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.375556 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.375615 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.375632 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.375659 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.375678 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.479050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.479101 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.479115 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.479132 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.479145 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.581968 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.582059 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.582086 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.582118 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.582144 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.685391 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.685439 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.685447 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.685463 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.685472 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.787552 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.787599 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.787609 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.787625 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.787637 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.837178 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:13 crc kubenswrapper[4867]: E1212 06:49:13.837323 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.890133 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.890181 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.890197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.890214 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.890247 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.993149 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.993195 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.993205 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.993247 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:13 crc kubenswrapper[4867]: I1212 06:49:13.993265 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:13Z","lastTransitionTime":"2025-12-12T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.096544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.096629 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.096655 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.096688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.096718 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.208264 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.208338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.208357 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.208385 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.208405 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.311911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.311986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.312004 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.312035 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.312055 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.355801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.355856 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.355866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.355888 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.355901 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.368986 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:14Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.373795 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.373833 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.373843 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.373860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.373872 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.386027 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:14Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.391235 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.391280 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.391293 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.391317 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.391331 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.406333 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:14Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.410369 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.410420 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.410436 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.410456 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.410467 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.426616 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:14Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.431076 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.431133 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.431145 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.431166 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.431179 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.444182 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:14Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.444339 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.445931 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.445979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.445989 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.446008 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.446018 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.549061 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.549132 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.549151 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.549180 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.549200 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.652038 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.652088 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.652100 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.652118 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.652128 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.754315 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.754366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.754378 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.754401 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.754412 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.838167 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.838287 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.838184 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.838479 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.838703 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:14 crc kubenswrapper[4867]: E1212 06:49:14.838906 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.856840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.856879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.856888 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.856904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.856914 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.959874 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.959959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.959978 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.959999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:14 crc kubenswrapper[4867]: I1212 06:49:14.960012 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:14Z","lastTransitionTime":"2025-12-12T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.063321 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.063426 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.063444 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.063468 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.063486 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.167194 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.167328 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.167348 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.167380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.167405 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.269881 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.270183 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.270215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.270276 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.270287 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.372866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.372915 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.372926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.372945 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.372960 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.477356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.477425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.477441 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.477473 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.477538 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.581116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.581170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.581187 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.581213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.581286 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.684922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.684981 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.684997 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.685026 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.685046 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.788263 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.788339 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.788356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.788384 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.788407 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.837326 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:15 crc kubenswrapper[4867]: E1212 06:49:15.837529 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.892392 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.892442 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.892460 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.892487 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.892508 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.995446 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.995508 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.995533 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.995564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:15 crc kubenswrapper[4867]: I1212 06:49:15.995587 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:15Z","lastTransitionTime":"2025-12-12T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.099270 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.099340 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.099363 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.099391 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.099411 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.202008 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.202074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.202093 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.202117 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.202135 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.305137 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.305197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.305221 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.305292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.305317 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.407811 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.407874 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.407888 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.407911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.407924 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.510900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.510959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.510975 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.510993 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.511003 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.614862 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.614926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.614942 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.614964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.614978 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.717608 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.717664 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.717679 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.717701 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.717714 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.820563 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.820604 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.820614 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.820633 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.820696 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.838137 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.838220 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:16 crc kubenswrapper[4867]: E1212 06:49:16.838303 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:16 crc kubenswrapper[4867]: E1212 06:49:16.838654 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.838775 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:16 crc kubenswrapper[4867]: E1212 06:49:16.838953 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.924203 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.924278 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.924296 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.924329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:16 crc kubenswrapper[4867]: I1212 06:49:16.924347 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:16Z","lastTransitionTime":"2025-12-12T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.027995 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.028052 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.028062 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.028078 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.028089 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.130891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.130968 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.130980 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.130999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.131013 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.235410 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.235486 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.235511 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.235547 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.235636 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.339322 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.339373 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.339383 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.339402 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.339424 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.442841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.442899 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.442917 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.442944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.442965 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.545974 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.546026 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.546039 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.546056 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.546069 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.650091 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.650161 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.650183 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.650211 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.650258 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.752423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.752475 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.752487 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.752505 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.752517 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.837321 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:17 crc kubenswrapper[4867]: E1212 06:49:17.837759 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.837974 4867 scope.go:117] "RemoveContainer" containerID="b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.855558 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.855590 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.855602 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.855620 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.855632 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.957961 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.958424 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.958523 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.958601 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:17 crc kubenswrapper[4867]: I1212 06:49:17.958668 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:17Z","lastTransitionTime":"2025-12-12T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.060548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.060579 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.060587 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.060600 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.060609 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.161008 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/1.log" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.165366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.165396 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.165407 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.165424 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.165435 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.168674 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.169463 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.184807 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.198201 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.220725 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.245006 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.269980 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.270037 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.270050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.270072 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.270090 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.271782 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.293886 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.332694 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.359709 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.372476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.372510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.372519 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.372534 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.372545 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.378625 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.392723 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.406545 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.417866 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.428604 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.442998 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.455292 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.469035 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.474632 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.474699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.474710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.474748 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.474761 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.486967 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:18Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.577310 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.577346 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.577356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.577375 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.577385 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.679986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.680022 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.680030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.680044 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.680054 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.782352 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.782406 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.782422 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.782444 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.782459 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.837957 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:18 crc kubenswrapper[4867]: E1212 06:49:18.838074 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.838148 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:18 crc kubenswrapper[4867]: E1212 06:49:18.838312 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.838408 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:18 crc kubenswrapper[4867]: E1212 06:49:18.838474 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.884298 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.884329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.884337 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.884349 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.884358 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.986994 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.987046 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.987062 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.987086 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:18 crc kubenswrapper[4867]: I1212 06:49:18.987104 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:18Z","lastTransitionTime":"2025-12-12T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.089679 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.089720 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.089731 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.089747 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.089757 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.193067 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.193154 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.193182 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.193297 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.193387 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.296408 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.296448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.296458 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.296472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.296481 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.399943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.399995 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.400011 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.400030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.400043 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.503823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.503884 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.503900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.503922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.503950 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.608073 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.608136 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.608147 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.608168 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.608180 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.711521 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.711567 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.711580 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.711598 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.711610 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.815510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.815556 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.815571 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.815591 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.815604 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.837133 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:19 crc kubenswrapper[4867]: E1212 06:49:19.837404 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.918705 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.918753 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.918765 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.918784 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:19 crc kubenswrapper[4867]: I1212 06:49:19.918796 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:19Z","lastTransitionTime":"2025-12-12T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.021752 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.021810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.021828 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.021853 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.021870 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.125173 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.125330 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.125354 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.125381 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.125406 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.181760 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/2.log" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.182880 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/1.log" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.187859 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d" exitCode=1 Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.187921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.187984 4867 scope.go:117] "RemoveContainer" containerID="b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.189391 4867 scope.go:117] "RemoveContainer" containerID="54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d" Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.189673 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.211466 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.228415 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.228457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.228471 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.228489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.228505 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.238372 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.252914 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.285887 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.298938 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.314974 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.331487 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.331546 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.331558 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.331576 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.331596 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.334084 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.355650 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.369918 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.387819 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.402176 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.411816 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.426896 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.434197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.434294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.434309 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.434332 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.434349 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.457096 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.469803 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.470199 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.470355 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.470401 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:36.470386495 +0000 UTC m=+64.041767764 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.490347 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.509492 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:20Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.537334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.537416 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.537438 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.537469 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.537492 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.641159 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.641209 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.641243 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.641263 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.641276 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.744644 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.744710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.744728 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.744753 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.744772 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.837689 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.837727 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.837856 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.838045 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.838216 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:20 crc kubenswrapper[4867]: E1212 06:49:20.838485 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.848006 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.848047 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.848059 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.848076 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.848091 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.951259 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.951342 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.951368 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.951404 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:20 crc kubenswrapper[4867]: I1212 06:49:20.951431 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:20Z","lastTransitionTime":"2025-12-12T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.056737 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.057431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.057503 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.057540 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.057562 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.160796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.160877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.160896 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.160927 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.160949 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.193644 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/2.log" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.264033 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.264084 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.264096 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.264120 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.264134 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.366587 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.367018 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.367085 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.367119 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.367141 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.469692 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.469787 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.469807 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.469841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.469869 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.573327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.573361 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.573370 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.573384 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.573394 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.675816 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.675881 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.675895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.675915 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.675932 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.779955 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.780007 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.780020 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.780041 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.780054 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.838264 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:21 crc kubenswrapper[4867]: E1212 06:49:21.838989 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.882561 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.882593 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.882605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.882619 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.882628 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.986074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.986111 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.986122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.986138 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:21 crc kubenswrapper[4867]: I1212 06:49:21.986149 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:21Z","lastTransitionTime":"2025-12-12T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.089609 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.089687 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.089699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.089725 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.089738 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.193331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.193415 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.193438 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.193476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.193499 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.296789 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.296894 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.296971 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.297010 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.297034 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.399617 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.399656 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.399665 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.399682 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.399692 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.493645 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.493898 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.493928 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.493938 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.493999 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:54.493984965 +0000 UTC m=+82.065366234 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.501986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.502014 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.502022 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.502035 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.502044 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.608054 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.608278 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.608324 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.608344 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.608355 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.696465 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.696625 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.696673 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.696730 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.696865 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.696934 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:54.69691283 +0000 UTC m=+82.268294129 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697211 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:49:54.697195206 +0000 UTC m=+82.268576515 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697315 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697360 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:54.697348459 +0000 UTC m=+82.268729768 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697447 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697481 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697495 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.697551 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:49:54.697536244 +0000 UTC m=+82.268917513 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.710943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.710976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.710985 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.711001 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.711012 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.812517 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.812554 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.812564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.812578 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.812590 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.837745 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.837795 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.837856 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.837943 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.838376 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:22 crc kubenswrapper[4867]: E1212 06:49:22.838461 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.852351 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.864375 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.874330 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.884541 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.894651 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.905097 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.913875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.913917 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.913929 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.913947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.913962 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:22Z","lastTransitionTime":"2025-12-12T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.923346 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.935906 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.944645 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.961605 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.971336 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.982770 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:22 crc kubenswrapper[4867]: I1212 06:49:22.992791 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:22Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.003964 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:23Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.014738 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:23Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.016037 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.016469 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.016502 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.016520 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.016532 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.025704 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:23Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.038537 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:23Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.118918 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.118964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.118976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.118994 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.119003 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.222470 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.222503 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.222514 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.222549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.222561 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.325065 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.325099 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.325108 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.325122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.325131 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.427991 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.428041 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.428053 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.428075 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.428093 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.530804 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.530847 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.530860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.530877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.530887 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.633832 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.633880 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.633892 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.633910 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.633925 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.736217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.736304 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.736313 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.736325 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.736334 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.837525 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:23 crc kubenswrapper[4867]: E1212 06:49:23.837691 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.838564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.838613 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.838630 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.838650 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.838667 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.941717 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.941784 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.941805 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.941835 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:23 crc kubenswrapper[4867]: I1212 06:49:23.941861 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:23Z","lastTransitionTime":"2025-12-12T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.044360 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.044423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.044437 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.044456 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.044467 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.147738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.147855 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.147868 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.147886 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.147898 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.250012 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.250052 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.250063 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.250078 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.250087 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.352943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.353018 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.353041 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.353074 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.353098 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.456076 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.456125 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.456136 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.456154 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.456168 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.559885 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.559965 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.559983 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.560019 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.560038 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.663094 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.663156 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.663182 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.663213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.663276 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.735590 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.735618 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.735626 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.735639 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.735649 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.754938 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.759721 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.759761 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.759775 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.759791 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.759801 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.773701 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.777826 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.777938 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.778049 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.778145 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.778245 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.791667 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.795329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.795366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.795378 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.795395 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.795408 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.808696 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.812442 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.812476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.812489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.812508 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.812517 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.825570 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:24Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.825714 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.827129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.827152 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.827162 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.827180 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.827214 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.837676 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.837954 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.837732 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.837697 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.838377 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:24 crc kubenswrapper[4867]: E1212 06:49:24.838459 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.929962 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.929991 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.929999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.930013 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:24 crc kubenswrapper[4867]: I1212 06:49:24.930022 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:24Z","lastTransitionTime":"2025-12-12T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.036596 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.036904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.036972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.037042 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.037098 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.139571 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.139626 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.139638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.139654 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.139666 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.243331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.243427 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.243448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.243479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.243501 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.347999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.348097 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.348116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.348144 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.348161 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.451949 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.452208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.452301 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.452373 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.452431 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.554926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.554968 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.554979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.554996 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.555008 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.658119 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.658381 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.658546 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.658656 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.658731 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.762616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.762688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.762709 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.762739 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.762760 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.837656 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:25 crc kubenswrapper[4867]: E1212 06:49:25.837798 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.866699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.866782 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.866797 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.866819 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.866835 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.970637 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.970724 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.970743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.970773 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:25 crc kubenswrapper[4867]: I1212 06:49:25.970791 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:25Z","lastTransitionTime":"2025-12-12T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.073882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.073926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.073937 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.073955 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.073968 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.177787 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.178380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.178524 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.178710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.178935 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.282756 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.282871 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.282908 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.282948 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.282980 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.386801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.386869 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.386885 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.386911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.386933 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.489672 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.490133 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.490250 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.490351 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.490468 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.593742 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.594181 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.594407 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.594573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.594692 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.703425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.703522 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.703549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.703585 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.703610 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.806520 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.806609 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.806632 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.806668 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.806690 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.837897 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.838119 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:26 crc kubenswrapper[4867]: E1212 06:49:26.838333 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.838393 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:26 crc kubenswrapper[4867]: E1212 06:49:26.838576 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:26 crc kubenswrapper[4867]: E1212 06:49:26.838724 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.909306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.909346 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.909355 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.909369 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:26 crc kubenswrapper[4867]: I1212 06:49:26.909378 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:26Z","lastTransitionTime":"2025-12-12T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.012067 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.012153 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.012174 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.012206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.012283 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.115350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.115685 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.115805 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.115917 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.116048 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.151309 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.163453 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.167537 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.194623 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.209561 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.218400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.218449 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.218463 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.218491 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.218505 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.225829 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.240418 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.258489 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.277339 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.298330 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.309610 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.320945 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.321122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.321245 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.321340 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.321429 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.336204 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.349644 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.364406 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.376769 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.390066 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.400163 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.410839 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.421658 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:27Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.422970 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.423150 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.423448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.423669 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.423865 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.526616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.526822 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.526910 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.526983 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.527044 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.629534 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.629812 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.629889 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.629958 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.630027 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.732472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.732523 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.732534 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.732555 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.732567 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.835573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.835636 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.835644 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.835661 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.835671 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.837825 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:27 crc kubenswrapper[4867]: E1212 06:49:27.838393 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.939013 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.939070 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.939083 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.939098 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:27 crc kubenswrapper[4867]: I1212 06:49:27.939108 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:27Z","lastTransitionTime":"2025-12-12T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.043692 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.043752 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.043767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.043790 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.043804 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.148811 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.148876 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.148895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.148924 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.148945 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.251007 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.251327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.251423 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.251510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.251579 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.355209 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.355330 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.355349 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.355382 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.355407 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.458031 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.458405 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.458533 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.458647 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.458772 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.563863 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.563954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.563974 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.564012 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.564053 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.666685 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.666725 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.666733 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.666769 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.666778 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.775582 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.775832 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.775930 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.776041 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.776122 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.864707 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:28 crc kubenswrapper[4867]: E1212 06:49:28.866158 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.864827 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.864771 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:28 crc kubenswrapper[4867]: E1212 06:49:28.866852 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:28 crc kubenswrapper[4867]: E1212 06:49:28.867142 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.878854 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.878901 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.878911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.878926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.878937 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.981955 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.981997 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.982008 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.982024 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:28 crc kubenswrapper[4867]: I1212 06:49:28.982035 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:28Z","lastTransitionTime":"2025-12-12T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.087594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.087656 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.087676 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.087705 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.087725 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.194142 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.194709 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.195076 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.195465 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.195707 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.299293 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.299865 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.300037 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.300187 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.300356 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.404283 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.404354 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.404373 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.404400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.404420 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.508839 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.508913 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.508940 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.508972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.508995 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.613780 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.613862 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.613881 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.613916 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.613935 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.717044 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.717114 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.717134 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.717157 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.717177 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.820334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.820389 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.820407 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.820431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.820448 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.837831 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:29 crc kubenswrapper[4867]: E1212 06:49:29.838065 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.925036 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.925121 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.925148 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.925185 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:29 crc kubenswrapper[4867]: I1212 06:49:29.925215 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:29Z","lastTransitionTime":"2025-12-12T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.029334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.029406 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.029426 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.029456 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.029482 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.133891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.134000 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.134019 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.134046 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.134067 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.237440 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.237563 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.237589 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.237620 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.237641 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.343280 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.343335 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.343348 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.343367 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.343379 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.446865 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.446918 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.446926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.446944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.446952 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.550325 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.550388 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.550400 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.550417 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.550427 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.653425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.653468 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.653476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.653490 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.653499 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.756796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.756867 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.756886 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.756909 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.756927 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.837453 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:30 crc kubenswrapper[4867]: E1212 06:49:30.837588 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.837756 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:30 crc kubenswrapper[4867]: E1212 06:49:30.837800 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.837894 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:30 crc kubenswrapper[4867]: E1212 06:49:30.837944 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.859688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.859737 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.859750 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.859768 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.859781 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.962068 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.962147 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.962163 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.962188 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:30 crc kubenswrapper[4867]: I1212 06:49:30.962209 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:30Z","lastTransitionTime":"2025-12-12T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.065407 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.065474 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.065496 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.065523 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.065544 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.168175 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.168199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.168206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.168219 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.168244 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.270566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.270785 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.270841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.270898 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.270958 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.373615 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.373857 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.373934 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.373999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.374061 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.476878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.477155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.477288 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.477386 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.477473 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.580694 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.580930 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.580991 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.581095 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.581768 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.685685 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.685758 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.685770 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.685789 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.685801 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.789639 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.789692 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.789703 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.789719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.789729 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.837775 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:31 crc kubenswrapper[4867]: E1212 06:49:31.838040 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.892683 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.892716 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.892725 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.892738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.892749 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.996901 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.996948 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.996958 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.996976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:31 crc kubenswrapper[4867]: I1212 06:49:31.996986 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:31Z","lastTransitionTime":"2025-12-12T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.100408 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.100488 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.100512 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.100544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.100567 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.202629 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.202678 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.202690 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.202706 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.202718 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.305305 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.305345 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.305361 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.305382 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.305398 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.408369 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.408405 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.408414 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.408427 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.408436 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.512352 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.512624 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.512644 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.512676 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.512699 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.614939 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.615020 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.615034 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.615052 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.615062 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.717311 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.717356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.717367 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.717381 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.717391 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.820258 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.820306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.820316 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.820335 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.820347 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.837657 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.837661 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:32 crc kubenswrapper[4867]: E1212 06:49:32.837909 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:32 crc kubenswrapper[4867]: E1212 06:49:32.838028 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.837758 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:32 crc kubenswrapper[4867]: E1212 06:49:32.838266 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.850524 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.861652 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.874376 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.884368 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.895256 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.906179 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.919841 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.922346 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.922380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.922390 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.922405 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.922416 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:32Z","lastTransitionTime":"2025-12-12T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.932667 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.942169 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.955685 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.968399 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.985408 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e71bf2ae65996c27af0c3327766c43ac4df415398c38a3942d83470742742b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:02Z\\\",\\\"message\\\":\\\"190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:01.879313 6304 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:01.879328 6304 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1212 06:49:01.879332 6304 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1212 06:49:01.879347 6304 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1212 06:49:01.879367 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:01.879415 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:01.879433 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:01.879434 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1212 06:49:01.879445 6304 factory.go:656] Stopping watch factory\\\\nI1212 06:49:01.879447 6304 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1212 06:49:01.879471 6304 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:01.879798 6304 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1212 06:49:01.879924 6304 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1212 06:49:01.879977 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:01.880014 6304 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:01.880101 6304 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:32 crc kubenswrapper[4867]: I1212 06:49:32.994865 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:32Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.010895 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.022374 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.024301 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.024334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.024349 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.024366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.024378 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.033466 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.046730 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.056349 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:33Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.126890 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.126961 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.126981 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.127000 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.127013 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.229712 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.229743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.229751 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.229764 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.229774 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.331868 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.331902 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.331911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.331926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.331936 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.434195 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.434242 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.434254 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.434271 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.434280 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.537273 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.537320 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.537334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.537350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.537362 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.640280 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.640320 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.640336 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.640355 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.640372 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.742841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.743256 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.743275 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.743298 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.743307 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.838075 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:33 crc kubenswrapper[4867]: E1212 06:49:33.838200 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.845722 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.845891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.845965 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.846031 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.846097 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.948825 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.949084 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.949184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.949285 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:33 crc kubenswrapper[4867]: I1212 06:49:33.949366 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:33Z","lastTransitionTime":"2025-12-12T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.051940 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.051979 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.051990 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.052007 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.052018 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.154641 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.154738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.154758 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.154781 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.154798 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.257740 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.257801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.257818 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.257843 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.257865 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.361005 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.361070 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.361089 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.361114 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.361131 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.464208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.464289 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.464306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.464328 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.464345 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.567431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.567473 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.567489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.567507 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.567523 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.669902 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.669965 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.669982 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.670002 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.670016 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.772536 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.772605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.772621 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.772637 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.772647 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.837768 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.837833 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.838306 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:34 crc kubenswrapper[4867]: E1212 06:49:34.838377 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:34 crc kubenswrapper[4867]: E1212 06:49:34.838300 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:34 crc kubenswrapper[4867]: E1212 06:49:34.838541 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.838823 4867 scope.go:117] "RemoveContainer" containerID="54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d" Dec 12 06:49:34 crc kubenswrapper[4867]: E1212 06:49:34.839068 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.864160 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.875092 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.876430 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.876463 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.876473 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.876490 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.876502 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.888634 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.904198 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.918976 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.932856 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.946322 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.959488 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972497 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972834 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972861 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972874 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:34 crc kubenswrapper[4867]: I1212 06:49:34.972902 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:34Z","lastTransitionTime":"2025-12-12T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.003443 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:34Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.010147 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.010207 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.010223 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.010280 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.010298 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.018950 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.031963 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.034512 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.035259 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.035280 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.035289 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.035304 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.035317 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.046754 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.047098 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.052110 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.052137 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.052147 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.052162 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.052174 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.057596 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.064994 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067377 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067622 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067647 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067655 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067667 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.067676 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.078554 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.081531 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.081653 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.083163 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.083181 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.083189 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.083202 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.083211 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.089787 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.108189 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.117924 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:35Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.185918 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.185953 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.185965 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.185980 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.185991 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.288640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.288668 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.288677 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.288689 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.288698 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.391357 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.391395 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.391425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.391438 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.391446 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.493976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.494035 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.494045 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.494061 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.494072 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.596324 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.596367 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.596379 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.596394 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.596410 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.699197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.699252 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.699261 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.699276 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.699287 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.802383 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.802999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.803208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.803743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.804053 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.837925 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:35 crc kubenswrapper[4867]: E1212 06:49:35.838280 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.907051 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.907092 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.907100 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.907114 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:35 crc kubenswrapper[4867]: I1212 06:49:35.907122 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:35Z","lastTransitionTime":"2025-12-12T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.009403 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.009665 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.009762 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.009863 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.009953 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.112960 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.113016 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.113025 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.113040 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.113050 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.220057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.220104 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.220424 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.220448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.220461 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.323128 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.323184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.323197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.323219 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.323253 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.425114 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.425155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.425163 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.425176 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.425185 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.527158 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.527215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.527261 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.527283 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.527298 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.550748 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:36 crc kubenswrapper[4867]: E1212 06:49:36.550943 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:36 crc kubenswrapper[4867]: E1212 06:49:36.551019 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:50:08.550999557 +0000 UTC m=+96.122380836 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.630057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.630087 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.630099 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.630115 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.630128 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.732139 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.732179 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.732194 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.732212 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.732252 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.834073 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.834109 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.834120 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.834136 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.834148 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.840289 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:36 crc kubenswrapper[4867]: E1212 06:49:36.840409 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.840457 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.840478 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:36 crc kubenswrapper[4867]: E1212 06:49:36.840648 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:36 crc kubenswrapper[4867]: E1212 06:49:36.840745 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.937023 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.937053 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.937061 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.937075 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:36 crc kubenswrapper[4867]: I1212 06:49:36.937084 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:36Z","lastTransitionTime":"2025-12-12T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.039493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.039530 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.039539 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.039554 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.039564 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.141706 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.141744 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.141752 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.141767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.141778 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.244807 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.244846 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.244854 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.244868 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.244878 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.247598 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/0.log" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.247644 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1918c53-1a6d-4317-a320-2e301b5a9db7" containerID="e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd" exitCode=1 Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.247671 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerDied","Data":"e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.248025 4867 scope.go:117] "RemoveContainer" containerID="e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.267674 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.279277 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.295758 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.311116 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.321702 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.333945 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.345544 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.347818 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.347848 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.347858 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.347872 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.347882 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.359929 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.374535 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.388059 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.403284 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.417934 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.430405 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.440560 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.450482 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.450523 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.450531 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.450548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.450560 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.451218 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.463763 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.482078 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.494269 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:37Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.552766 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.552794 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.552803 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.552815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.552824 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.655172 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.655206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.655215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.655241 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.655250 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.757123 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.757176 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.757187 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.757204 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.757216 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.837087 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:37 crc kubenswrapper[4867]: E1212 06:49:37.837203 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.862410 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.862469 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.862499 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.862526 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.862593 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.965683 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.965728 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.965739 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.965753 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:37 crc kubenswrapper[4867]: I1212 06:49:37.965764 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:37Z","lastTransitionTime":"2025-12-12T06:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.068701 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.068768 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.068781 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.068803 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.068817 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.172166 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.172194 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.172204 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.172217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.172242 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.252776 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/0.log" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.252841 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerStarted","Data":"9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.266628 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.276292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.276343 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.276351 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.276365 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.276376 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.280092 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.294927 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.305950 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.318065 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.328139 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.339801 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.351632 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.359836 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.372810 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.378599 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.378631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.378638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.378653 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.378661 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.389411 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.406613 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.415974 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.433708 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.443328 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.454563 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.467593 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.476991 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:38Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.480833 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.480868 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.480880 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.480896 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.480907 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.583279 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.583343 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.583356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.583372 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.583387 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.686014 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.686050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.686058 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.686071 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.686081 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.788281 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.788319 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.788329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.788344 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.788353 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.837861 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.837894 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.837879 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:38 crc kubenswrapper[4867]: E1212 06:49:38.837999 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:38 crc kubenswrapper[4867]: E1212 06:49:38.838312 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:38 crc kubenswrapper[4867]: E1212 06:49:38.838214 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.890802 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.890839 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.890850 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.890868 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.890883 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.993047 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.993089 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.993101 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.993118 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:38 crc kubenswrapper[4867]: I1212 06:49:38.993129 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:38Z","lastTransitionTime":"2025-12-12T06:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.095685 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.095719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.095728 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.095741 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.095751 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.197366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.197402 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.197413 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.197427 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.197438 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.298958 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.298997 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.299012 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.299028 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.299039 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.401277 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.401319 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.401329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.401344 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.401354 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.503008 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.503053 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.503065 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.503081 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.503092 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.604924 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.604963 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.604971 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.604985 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.604995 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.707623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.707667 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.707677 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.707693 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.707703 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.810072 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.810112 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.810122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.810137 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.810148 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.837603 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:39 crc kubenswrapper[4867]: E1212 06:49:39.837742 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.912566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.912603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.912612 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.912629 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:39 crc kubenswrapper[4867]: I1212 06:49:39.912641 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:39Z","lastTransitionTime":"2025-12-12T06:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.014724 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.014765 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.014775 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.014799 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.014809 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.116815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.116852 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.116861 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.116876 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.116887 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.219051 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.219121 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.219131 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.219144 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.219153 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.321126 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.321168 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.321179 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.321197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.321209 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.423638 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.423691 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.423700 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.423713 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.423721 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.526029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.526095 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.526122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.526153 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.526177 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.628188 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.628245 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.628259 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.628274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.628285 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.730980 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.731046 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.731070 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.731101 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.731128 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.833808 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.833878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.833898 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.833925 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.833946 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.838106 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.838292 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:40 crc kubenswrapper[4867]: E1212 06:49:40.838371 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.838395 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:40 crc kubenswrapper[4867]: E1212 06:49:40.838742 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:40 crc kubenswrapper[4867]: E1212 06:49:40.838796 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.935914 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.935959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.935972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.935996 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:40 crc kubenswrapper[4867]: I1212 06:49:40.936010 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:40Z","lastTransitionTime":"2025-12-12T06:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.038364 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.038403 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.038412 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.038425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.038434 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.140639 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.140690 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.140700 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.140717 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.140728 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.243293 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.243339 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.243350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.243367 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.243378 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.345859 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.345897 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.345906 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.345922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.345932 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.447573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.447643 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.447662 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.447714 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.447736 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.550698 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.550742 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.550753 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.550770 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.550782 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.656860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.656899 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.656909 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.656925 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.656937 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.760181 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.760250 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.760262 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.760277 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.760291 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.837313 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:41 crc kubenswrapper[4867]: E1212 06:49:41.837471 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.863419 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.863445 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.863454 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.863467 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.863480 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.965564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.965616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.965627 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.965644 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:41 crc kubenswrapper[4867]: I1212 06:49:41.965656 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:41Z","lastTransitionTime":"2025-12-12T06:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.067890 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.067937 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.067947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.067962 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.067973 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.170982 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.171021 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.171029 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.171045 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.171053 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.272922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.272976 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.272987 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.273005 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.273015 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.375318 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.375350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.375360 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.375376 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.375389 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.477431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.477471 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.477479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.477496 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.477504 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.579783 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.579829 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.579840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.579858 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.579872 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.682800 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.682833 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.682844 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.682860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.682872 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.785085 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.785130 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.785142 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.785166 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.785177 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.837906 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.837922 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:42 crc kubenswrapper[4867]: E1212 06:49:42.838040 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.838113 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:42 crc kubenswrapper[4867]: E1212 06:49:42.838213 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:42 crc kubenswrapper[4867]: E1212 06:49:42.838328 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.849267 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.860195 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.868864 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.880163 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.889283 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.889331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.889339 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.889353 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.889363 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.891334 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.901685 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.914839 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.929063 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.939396 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.951108 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.962322 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.982158 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.991849 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.991890 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.991899 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.991912 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.991923 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:42Z","lastTransitionTime":"2025-12-12T06:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:42 crc kubenswrapper[4867]: I1212 06:49:42.993891 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:42Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.010767 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.019347 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.028936 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.041808 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.051073 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:43Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.093841 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.093875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.093884 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.093898 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.093908 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.196833 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.196900 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.196911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.196927 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.196938 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.299719 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.299751 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.299759 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.299771 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.299782 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.402354 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.402396 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.402404 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.402416 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.402425 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.504892 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.504931 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.504941 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.504954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.504963 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.607527 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.607576 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.607586 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.607603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.607616 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.709666 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.709700 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.709710 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.709723 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.709734 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.812597 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.812641 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.812652 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.812668 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.812681 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.837272 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:43 crc kubenswrapper[4867]: E1212 06:49:43.837432 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.914525 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.914565 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.914578 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.914595 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:43 crc kubenswrapper[4867]: I1212 06:49:43.914609 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:43Z","lastTransitionTime":"2025-12-12T06:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.017866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.017906 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.017915 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.017931 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.017943 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.120170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.120215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.120249 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.120265 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.120277 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.222772 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.222801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.222809 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.222821 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.222830 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.325443 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.325477 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.325487 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.325504 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.325518 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.428336 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.428396 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.428408 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.428426 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.428438 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.530640 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.530694 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.530707 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.530724 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.530737 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.633237 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.633276 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.633283 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.633297 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.633307 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.735153 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.735206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.735217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.735252 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.735265 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837166 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837210 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:44 crc kubenswrapper[4867]: E1212 06:49:44.837342 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837382 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837602 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837651 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: E1212 06:49:44.837642 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837676 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: E1212 06:49:44.837711 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837769 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.837834 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.939812 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.939862 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.939878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.939901 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:44 crc kubenswrapper[4867]: I1212 06:49:44.939917 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:44Z","lastTransitionTime":"2025-12-12T06:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.042552 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.042624 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.042641 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.042664 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.042689 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.145387 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.145467 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.145496 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.145527 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.145551 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.247838 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.247905 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.247922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.247947 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.247967 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.350140 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.350182 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.350193 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.350207 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.350218 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.452942 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.452986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.452999 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.453015 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.453026 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.454822 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.454878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.454893 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.454913 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.454933 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.469813 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.474909 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.474991 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.475005 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.475021 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.475032 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.488450 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.492743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.492773 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.492782 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.492796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.492806 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.509272 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.513261 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.513305 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.513320 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.513340 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.513355 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.526299 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.529506 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.529535 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.529543 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.529573 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.529584 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.540729 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:45Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.540900 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.555218 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.555279 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.555290 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.555306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.555318 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.657434 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.657471 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.657482 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.657512 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.657525 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.760292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.760324 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.760338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.760380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.760392 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.837094 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:45 crc kubenswrapper[4867]: E1212 06:49:45.837297 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.838192 4867 scope.go:117] "RemoveContainer" containerID="54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.862564 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.862844 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.862857 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.862875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.862887 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.964896 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.964932 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.964943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.964958 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:45 crc kubenswrapper[4867]: I1212 06:49:45.964969 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:45Z","lastTransitionTime":"2025-12-12T06:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.067312 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.067338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.067346 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.067359 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.067368 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.170167 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.170217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.170248 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.170265 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.170276 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.272442 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.272476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.272488 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.272503 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.272514 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.275975 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/2.log" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.278932 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.279494 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.304027 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.331463 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.340772 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.351815 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.363032 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.375158 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.375193 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.375204 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.375238 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.375252 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.379272 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.390002 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.401401 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.412300 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.423048 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.439042 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.450955 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.471333 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.477023 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.477050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.477060 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.477075 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.477088 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.484403 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.497398 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.509638 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.521311 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.530259 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:46Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.579718 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.579752 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.579763 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.579778 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.579789 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.682350 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.682407 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.682426 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.682454 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.682474 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.786959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.786998 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.787009 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.787026 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.787037 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.837583 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.837623 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.837658 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:46 crc kubenswrapper[4867]: E1212 06:49:46.837754 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:46 crc kubenswrapper[4867]: E1212 06:49:46.837821 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:46 crc kubenswrapper[4867]: E1212 06:49:46.837896 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.890160 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.890287 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.890310 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.890344 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.890364 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.992814 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.992857 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.992865 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.992878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:46 crc kubenswrapper[4867]: I1212 06:49:46.992887 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:46Z","lastTransitionTime":"2025-12-12T06:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.094822 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.094866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.094877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.094895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.094908 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.197433 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.197469 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.197479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.197493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.197506 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.283518 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/3.log" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.284275 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/2.log" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.287817 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" exitCode=1 Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.287871 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.287906 4867 scope.go:117] "RemoveContainer" containerID="54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.288508 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 06:49:47 crc kubenswrapper[4867]: E1212 06:49:47.288656 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.299548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.299595 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.299606 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.299623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.299636 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.303238 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.316449 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.325609 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.335078 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.348312 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.366935 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54bfde0e64d6e16bb53034aef03386358ab2a78f604ceb2b6ea9668bc519af7d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:19Z\\\",\\\"message\\\":\\\"ress service\\\\nI1212 06:49:18.829134 6513 egressservice_zone_endpointslice.go:80] Ignoring updating openshift-marketplace/redhat-operators for endpointslice openshift-marketplace/redhat-operators-jntmp as it is not a known egress service\\\\nI1212 06:49:18.827897 6513 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1212 06:49:18.829281 6513 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1212 06:49:18.829317 6513 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1212 06:49:18.829325 6513 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1212 06:49:18.829351 6513 factory.go:656] Stopping watch factory\\\\nI1212 06:49:18.829385 6513 handler.go:208] Removed *v1.Node event handler 7\\\\nI1212 06:49:18.829405 6513 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1212 06:49:18.829418 6513 handler.go:208] Removed *v1.Node event handler 2\\\\nI1212 06:49:18.828517 6513 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1212 06:49:18.830294 6513 ovnkube.go:599] Stopped ovnkube\\\\nI1212 06:49:18.830336 6513 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1212 06:49:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:46Z\\\",\\\"message\\\":\\\"te:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1212 06:49:46.682439 6906 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1212 06:49:46.682444 6906 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1212 06:49:46.682445 6906 services_controller.go:452] Built service openshift-operator-lifecycle-manager/olm-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682452 6906 services_controller.go:453] Built service openshift-operator-lifecycle-manager/olm-operator-metrics template LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682456 6906 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1212 06:49:46.682458 6906 services_controller.go:454] Service openshift-operator-lifecycle-manager/olm-operator-metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1212 06:49:46.682498 6906 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.379192 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.392099 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.401922 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.401982 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.401992 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.402006 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.402039 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.406156 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.418320 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.434551 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.447130 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.466635 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.479939 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.491496 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.504510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.504552 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.504563 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.504582 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.504595 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.505274 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.518685 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.531442 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:47Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.613213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.613279 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.613289 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.613306 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.613317 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.715537 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.715580 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.715593 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.715611 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.715624 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.818330 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.818373 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.818382 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.818397 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.818406 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.837694 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:47 crc kubenswrapper[4867]: E1212 06:49:47.837843 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.920753 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.920788 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.920796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.920809 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:47 crc kubenswrapper[4867]: I1212 06:49:47.920818 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:47Z","lastTransitionTime":"2025-12-12T06:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.022760 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.022790 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.022798 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.022811 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.022819 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.124939 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.124997 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.125010 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.125027 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.125039 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.227161 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.227199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.227209 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.227241 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.227252 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.291337 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/3.log" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.294655 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 06:49:48 crc kubenswrapper[4867]: E1212 06:49:48.294840 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.305497 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.315786 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.329281 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.329548 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.329616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.329682 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.329745 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.330014 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:46Z\\\",\\\"message\\\":\\\"te:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1212 06:49:46.682439 6906 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1212 06:49:46.682444 6906 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1212 06:49:46.682445 6906 services_controller.go:452] Built service openshift-operator-lifecycle-manager/olm-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682452 6906 services_controller.go:453] Built service openshift-operator-lifecycle-manager/olm-operator-metrics template LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682456 6906 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1212 06:49:46.682458 6906 services_controller.go:454] Service openshift-operator-lifecycle-manager/olm-operator-metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1212 06:49:46.682498 6906 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.338322 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.353336 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.360785 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.369625 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.381153 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.390449 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.399852 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.408277 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.416596 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.425943 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.432294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.432329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.432339 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.432354 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.432363 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.434471 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.443534 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.454553 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.463873 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.471011 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:48Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.535413 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.535461 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.535478 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.535497 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.535510 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.638068 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.638129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.638149 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.638180 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.638201 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.740419 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.740460 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.740471 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.740486 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.740495 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.837504 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.837502 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:48 crc kubenswrapper[4867]: E1212 06:49:48.837727 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.837818 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:48 crc kubenswrapper[4867]: E1212 06:49:48.837969 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:48 crc kubenswrapper[4867]: E1212 06:49:48.838077 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.844287 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.844360 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.844443 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.844532 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.844739 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.947824 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.947967 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.947985 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.948004 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:48 crc kubenswrapper[4867]: I1212 06:49:48.948015 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:48Z","lastTransitionTime":"2025-12-12T06:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.051329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.051376 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.051386 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.051401 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.051410 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.155580 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.155646 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.155673 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.155754 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.155781 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.257837 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.257867 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.257875 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.257888 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.257898 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.360745 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.360794 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.360805 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.360823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.360834 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.463031 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.463650 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.463720 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.463823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.463903 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.566429 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.566529 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.566549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.566574 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.566592 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.669320 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.669375 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.669399 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.669421 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.669435 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.772133 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.772185 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.772198 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.772217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.772247 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.837358 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:49 crc kubenswrapper[4867]: E1212 06:49:49.837480 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.875691 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.875964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.876049 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.876134 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.876201 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.978964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.979371 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.979537 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.979729 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:49 crc kubenswrapper[4867]: I1212 06:49:49.979935 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:49Z","lastTransitionTime":"2025-12-12T06:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.083597 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.083666 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.083690 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.083720 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.083743 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.186027 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.186366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.186509 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.186631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.186756 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.289709 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.289775 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.289796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.289823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.289845 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.392871 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.392932 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.392948 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.392971 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.392989 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.495860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.496492 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.496566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.496603 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.496625 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.599294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.599330 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.599338 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.599352 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.599361 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.701420 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.701460 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.701470 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.701485 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.701496 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.804176 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.804263 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.804279 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.804300 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.804319 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.837871 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.837905 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:50 crc kubenswrapper[4867]: E1212 06:49:50.838029 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.838088 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:50 crc kubenswrapper[4867]: E1212 06:49:50.838288 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:50 crc kubenswrapper[4867]: E1212 06:49:50.838368 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.906850 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.906893 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.906903 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.906920 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:50 crc kubenswrapper[4867]: I1212 06:49:50.906931 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:50Z","lastTransitionTime":"2025-12-12T06:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.009504 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.009538 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.009550 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.009565 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.009577 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.111538 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.111576 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.111605 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.111620 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.111628 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.214018 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.214062 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.214072 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.214089 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.214100 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.316695 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.316739 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.316751 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.316767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.316779 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.418881 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.418917 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.418925 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.418937 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.418946 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.521417 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.521472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.521484 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.521501 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.521515 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.624248 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.624285 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.624314 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.624328 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.624337 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.726763 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.726828 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.726846 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.726873 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.726890 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.830697 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.830747 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.830759 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.830780 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.830794 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.837140 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:51 crc kubenswrapper[4867]: E1212 06:49:51.837312 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.933966 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.934022 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.934033 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.934052 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:51 crc kubenswrapper[4867]: I1212 06:49:51.934065 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:51Z","lastTransitionTime":"2025-12-12T06:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.035971 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.036021 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.036031 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.036046 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.036056 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.138730 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.138777 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.138788 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.138803 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.138816 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.240780 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.240836 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.240846 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.240859 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.240869 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.343116 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.343157 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.343170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.343187 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.343198 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.446105 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.446155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.446167 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.446186 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.446199 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.548188 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.548235 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.548245 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.548260 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.548299 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.650308 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.650339 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.650348 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.650361 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.650370 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.751897 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.751975 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.751989 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.752003 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.752035 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.838084 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.838084 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:52 crc kubenswrapper[4867]: E1212 06:49:52.838391 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:52 crc kubenswrapper[4867]: E1212 06:49:52.838613 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.838628 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:52 crc kubenswrapper[4867]: E1212 06:49:52.838744 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.852518 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.853771 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.853804 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.853815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.853832 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.853845 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.865254 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57b521de706ed4468c174e8fe1f56d7fadda9c124c931b24fe162f14da02d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab77672caf26948cb24acd4cf959e3ff763964d7ea7d26e7469d3503096e6c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.883213 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f74e7557dbcff877f4ada53794a7c798fd497f79537d90f0d594758070e5e4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.898409 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z8lss" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1918c53-1a6d-4317-a320-2e301b5a9db7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:37Z\\\",\\\"message\\\":\\\"2025-12-12T06:48:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9\\\\n2025-12-12T06:48:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c7cc4abb-80bd-4e29-971a-8dd72e4923d9 to /host/opt/cni/bin/\\\\n2025-12-12T06:48:52Z [verbose] multus-daemon started\\\\n2025-12-12T06:48:52Z [verbose] Readiness Indicator file check\\\\n2025-12-12T06:49:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9tmjr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z8lss\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.909831 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d604431e-f3e6-483a-abcc-92c8c5a5ab82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://445cc8fb2087f0dc5e44d331d510b0ac2e7024a9e55095cff5306da82e862356\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e291496550b8ef4e2dc506d2903a710797c1f362f60bba2d2734bc7d94ce2e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:49:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f5hn9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vrtlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.925739 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d97449ef-66de-421e-a9d6-8b003d2cb2ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b1189da8aa7f60b3ad8d1391887d8c4fb82168961d9c6c2fb994afd7f8161a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://746f987f20f1fcd843dfe456afe3f603064747ad97b2ed8d6fbb0e531d3c5584\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28021f5e97246b5b80e85ded976903d96a68325aff97e59a096ac475de303019\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b18ea662be0d7d5e3d789ffaa1c11fbf99b68023ad03873b9a3cd5d986cc1bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.943400 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4cb29ed6-1017-4d04-a984-ba197f53ef44\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"ed_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1765522125\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1765522125\\\\\\\\\\\\\\\" (2025-12-12 05:48:45 +0000 UTC to 2026-12-12 05:48:45 +0000 UTC (now=2025-12-12 06:48:50.456785704 +0000 UTC))\\\\\\\"\\\\nI1212 06:48:50.456817 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1212 06:48:50.456840 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1212 06:48:50.457900 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1212 06:48:50.457956 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1212 06:48:50.458015 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1795339564/tls.crt::/tmp/serving-cert-1795339564/tls.key\\\\\\\"\\\\nI1212 06:48:50.458197 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1212 06:48:50.458751 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1212 06:48:50.458773 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1212 06:48:50.459087 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459103 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1212 06:48:50.459124 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1212 06:48:50.459130 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF1212 06:48:50.461126 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.957061 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.957103 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.957117 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.957135 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.957146 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:52Z","lastTransitionTime":"2025-12-12T06:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.959901 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.971672 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dsjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74c23654-e77b-459c-b6fe-aa39e7e0bed2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23976f9f6020c5976932fe5e2235ef1e7708686b367d91860a94e972a58ec2f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx7q9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dsjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.986407 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a1514b0-51f8-47d2-afa2-c13983f7dbde\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca13f012b3881484129cab17d713e072dfb1fedeab2c0b93531e36462871aea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://173db2092cb329e891f8740a1eb81fe743b9e17189b33ec906db2961d3856915\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d53f98a919d7005bd052270b538c0801e222b4c4fa6c39e401eb8ed03f5605b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:52 crc kubenswrapper[4867]: I1212 06:49:52.999657 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32852b13f0d4ef9f83e47edc034f95121726f41424e423c28457750c62749168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:52Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.022210 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f519f9a-7a2d-4193-8c21-ae7840036dae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-12T06:49:46Z\\\",\\\"message\\\":\\\"te:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1212 06:49:46.682439 6906 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1212 06:49:46.682444 6906 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1212 06:49:46.682445 6906 services_controller.go:452] Built service openshift-operator-lifecycle-manager/olm-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682452 6906 services_controller.go:453] Built service openshift-operator-lifecycle-manager/olm-operator-metrics template LB for network=default: []services.LB{}\\\\nI1212 06:49:46.682456 6906 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1212 06:49:46.682458 6906 services_controller.go:454] Service openshift-operator-lifecycle-manager/olm-operator-metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1212 06:49:46.682498 6906 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-12T06:49:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6mhrh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbsqz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.038456 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d854288fc85eae673eaabb4e9ca1d94c3fe37667ad9d71c578c5101851540488\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9rjk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-q8lbh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.060366 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.060411 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.060424 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.060441 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.060454 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.064594 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3e5efe2d-678a-4181-8d8b-eb6e001d3a49\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c560f3b155851bbf34af53d40e270d2d9ded9f48d769a02218aa611b029b33f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9bfac51d91983c921e45f2f9390fe33dfbf65f966bf8b4280efbce35ce65e64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6836942bb76e5fd68749dafc6f24a5909d97617ed49ad3fcdc00733fb0859c8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d8857a51bb5f8f13238ac84341555cd8de240fe7316a13a3196975211b8e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9825d3ab70fc413dc022bdc1d387242f79ca4f8c88c32b4d6350c5918eceea8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2cb2921dbc4d610a8e6c2d67336b1807c4ea35312b1d6a7e4b20d5bcb165d635\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ae9f56c63dc80e83d9b8d284f9fef75f9ed25f22367ff01f739ff278f1154f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e43c0af02e169c261466fa5b8162ec03d50da645d7f89b3c4e80afee61f92e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.080175 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-njfkw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c7878997-f37a-4633-ae33-391200ab7e05\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72e22a712d94deb213103c28fadb3b05c20eacbf46f6d95e6a0d4ab015b559d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8bsld\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-njfkw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.095745 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.116526 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8287969-549f-480e-90bd-4478730313af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f062bf6d3ee9b027a2cc9548db3093f85a1fed386c91e5ea213e64bf0f9df3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-12T06:48:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa09a3a381abb27548ffd9699e4f45ca15e5e542dcf56c9842d89903cd9f2a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ddb48f0bb1e71d310923e860eed7fc326a75640761c7e1a3c7a3c16c6ab7112\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://750c05e108ca029f73f74e97447d504270b7229eba2fd8ec7530e7afb9152dd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec23dcee27f7e85b1b680533ed9b9b7a346c6a1dc85ada6ea452a8b9c83be5df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fccab18d254e8175f9139c4ffd732757025c82a436aefa0c196872ddec660c6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8581314ce1691a94afa281c9ee64902230577ddf984ff61ea499516d895c706a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-12T06:48:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-12T06:48:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wdgnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:48:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9dpkr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.130951 4867 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssjqs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-12T06:49:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hj2lv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:53Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.163112 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.163165 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.163176 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.163199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.163211 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.266140 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.266197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.266212 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.266270 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.266296 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.368849 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.368912 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.368933 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.368957 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.368975 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.472823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.472898 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.472918 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.472948 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.472973 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.575790 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.575824 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.575832 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.575848 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.575859 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.679797 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.679861 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.679874 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.679897 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.679907 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.782741 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.782801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.782820 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.782845 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.782862 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.837204 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:53 crc kubenswrapper[4867]: E1212 06:49:53.837461 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.885783 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.885859 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.885882 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.885927 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.885951 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.988837 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.988908 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.988925 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.988949 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:53 crc kubenswrapper[4867]: I1212 06:49:53.988968 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:53Z","lastTransitionTime":"2025-12-12T06:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.092661 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.092714 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.092725 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.092742 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.092753 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.196606 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.196676 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.196699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.196728 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.196753 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.299458 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.299494 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.299503 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.299516 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.299524 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.401666 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.401711 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.401721 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.401739 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.401752 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.504100 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.504158 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.504175 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.504199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.504254 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.533768 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.534002 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.534041 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.534061 4867 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.534139 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.534114292 +0000 UTC m=+146.105495601 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.606880 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.606953 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.606972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.606996 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.607012 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.710057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.710143 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.710165 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.710193 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.710216 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.736814 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.736982 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.736952414 +0000 UTC m=+146.308333713 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.737078 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.737185 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.737286 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737356 4867 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737440 4867 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737466 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.737437217 +0000 UTC m=+146.308818526 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737488 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737507 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.737489168 +0000 UTC m=+146.308870467 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737524 4867 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737548 4867 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.737677 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.737653813 +0000 UTC m=+146.309035122 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.812756 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.812860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.812878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.812896 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.812908 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.837670 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.837718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.837842 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.837921 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.838047 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:54 crc kubenswrapper[4867]: E1212 06:49:54.838126 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.917513 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.917591 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.917616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.917646 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:54 crc kubenswrapper[4867]: I1212 06:49:54.917668 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:54Z","lastTransitionTime":"2025-12-12T06:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.021743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.022125 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.022135 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.022151 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.022161 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.125422 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.125495 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.125516 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.125544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.125566 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.228199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.228385 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.228772 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.229164 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.229548 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.333088 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.333129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.333138 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.333155 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.333168 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.436444 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.436490 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.436501 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.436520 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.436531 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.540086 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.540219 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.540255 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.540276 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.540290 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.642778 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.642826 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.642840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.642861 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.642875 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.645129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.645170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.645184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.645199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.645210 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.662796 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.667079 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.667122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.667131 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.667146 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.667157 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.681057 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.684948 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.684998 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.685013 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.685032 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.685046 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.700861 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.705139 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.705184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.705196 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.705214 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.705249 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.721033 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.726205 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.726312 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.726332 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.726362 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.726384 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.743902 4867 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-12T06:49:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b0f6ee30-6d15-417f-bdda-47a426b4f903\\\",\\\"systemUUID\\\":\\\"3195cc84-39c8-4f61-8d51-1e423683d247\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-12T06:49:55Z is after 2025-08-24T17:21:41Z" Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.744063 4867 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.745987 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.746020 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.746032 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.746050 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.746064 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.837496 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:55 crc kubenswrapper[4867]: E1212 06:49:55.838210 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.849199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.849284 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.849302 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.849327 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.849348 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.952800 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.952849 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.952862 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.952878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:55 crc kubenswrapper[4867]: I1212 06:49:55.952890 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:55Z","lastTransitionTime":"2025-12-12T06:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.055162 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.055214 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.055264 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.055289 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.055306 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.158106 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.158144 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.158156 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.158172 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.158182 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.261053 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.261112 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.261121 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.261135 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.261145 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.363879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.363916 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.363927 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.363946 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.363957 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.466926 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.466967 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.466977 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.466990 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.466999 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.569914 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.569998 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.570016 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.570048 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.570068 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.672908 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.672944 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.672954 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.672968 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.672980 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.775483 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.775539 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.775551 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.775569 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.775581 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.837584 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.837630 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.837608 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:56 crc kubenswrapper[4867]: E1212 06:49:56.837753 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:56 crc kubenswrapper[4867]: E1212 06:49:56.837822 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:56 crc kubenswrapper[4867]: E1212 06:49:56.837936 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.878125 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.878183 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.878199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.878252 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.878272 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.981377 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.981422 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.981434 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.981451 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:56 crc kubenswrapper[4867]: I1212 06:49:56.981464 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:56Z","lastTransitionTime":"2025-12-12T06:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.083830 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.083867 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.083877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.083891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.083903 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.186020 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.186057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.186065 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.186082 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.186093 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.288006 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.288066 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.288081 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.288106 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.288173 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.391199 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.391272 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.391284 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.391303 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.391464 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.494154 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.494257 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.494270 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.494331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.494346 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.596691 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.596730 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.596743 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.596760 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.596772 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.699134 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.699174 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.699184 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.699201 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.699215 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.801653 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.801699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.801711 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.801730 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.801743 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.837513 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:57 crc kubenswrapper[4867]: E1212 06:49:57.837743 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.904498 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.904570 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.904594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.904623 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:57 crc kubenswrapper[4867]: I1212 06:49:57.904647 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:57Z","lastTransitionTime":"2025-12-12T06:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.008163 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.008217 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.008243 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.008263 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.008275 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.110972 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.111006 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.111014 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.111044 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.111054 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.214474 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.214522 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.214532 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.214551 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.214563 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.317878 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.317929 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.317943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.317963 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.317976 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.420760 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.420866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.420895 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.420931 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.420955 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.523334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.523427 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.523458 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.523495 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.523517 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.626001 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.626049 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.626060 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.626076 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.626087 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.728959 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.729066 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.729092 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.729123 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.729153 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.831650 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.831688 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.831699 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.831716 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.831727 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.838075 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.838086 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:49:58 crc kubenswrapper[4867]: E1212 06:49:58.838173 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.838220 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:49:58 crc kubenswrapper[4867]: E1212 06:49:58.838334 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:49:58 crc kubenswrapper[4867]: E1212 06:49:58.838381 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.934274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.934310 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.934318 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.934331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:58 crc kubenswrapper[4867]: I1212 06:49:58.934340 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:58Z","lastTransitionTime":"2025-12-12T06:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.036497 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.036525 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.036533 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.036563 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.036573 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.140767 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.140810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.140824 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.140840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.140850 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.243219 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.243324 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.243347 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.243375 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.243395 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.345879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.345932 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.345946 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.345967 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.345983 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.448907 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.448964 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.448981 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.449007 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.449024 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.551733 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.551798 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.551813 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.551837 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.551855 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.654282 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.654325 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.654334 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.654349 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.654362 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.757904 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.757955 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.757969 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.757986 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.758006 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.837885 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:49:59 crc kubenswrapper[4867]: E1212 06:49:59.838546 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.861213 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.861282 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.861294 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.861310 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.861322 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.964484 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.964549 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.964572 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.964596 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:49:59 crc kubenswrapper[4867]: I1212 06:49:59.964613 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:49:59Z","lastTransitionTime":"2025-12-12T06:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.067428 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.067476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.067489 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.067508 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.067520 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.170480 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.170858 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.171080 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.171356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.171768 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.274208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.274287 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.274300 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.274321 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.274334 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.377129 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.377186 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.377197 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.377218 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.377249 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.480097 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.480206 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.480215 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.480258 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.480270 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.582695 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.582809 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.582840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.582870 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.582887 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.686296 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.686345 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.686362 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.686386 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.686404 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.789292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.789331 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.789356 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.789370 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.789379 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.837595 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:00 crc kubenswrapper[4867]: E1212 06:50:00.837766 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.838037 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:00 crc kubenswrapper[4867]: E1212 06:50:00.838134 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.838522 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:00 crc kubenswrapper[4867]: E1212 06:50:00.838630 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.892164 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.892208 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.892239 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.892254 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.892264 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.995036 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.995073 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.995090 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.995106 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:00 crc kubenswrapper[4867]: I1212 06:50:00.995115 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:00Z","lastTransitionTime":"2025-12-12T06:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.097241 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.097274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.097301 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.097315 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.097323 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.199736 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.199781 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.199796 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.199815 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.199828 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.301784 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.301828 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.301840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.301857 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.301867 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.404241 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.404274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.404282 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.404296 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.404306 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.506842 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.506883 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.506891 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.506905 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.506915 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.608958 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.608995 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.609004 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.609016 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.609024 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.711345 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.711409 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.711431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.711459 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.711482 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.814558 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.814639 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.814651 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.814667 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.814680 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.837341 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:01 crc kubenswrapper[4867]: E1212 06:50:01.837507 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.838658 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 06:50:01 crc kubenswrapper[4867]: E1212 06:50:01.838926 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.918373 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.918431 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.918447 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.918472 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:01 crc kubenswrapper[4867]: I1212 06:50:01.918489 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:01Z","lastTransitionTime":"2025-12-12T06:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.021099 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.021131 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.021141 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.021154 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.021163 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.123631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.124007 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.124125 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.124249 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.124357 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.228387 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.228479 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.228508 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.228544 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.228578 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.332186 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.332263 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.332277 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.332292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.332300 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.435810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.435879 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.435897 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.436392 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.436453 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.539425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.539465 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.539476 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.539492 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.539505 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.642264 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.642305 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.642314 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.642329 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.642337 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.744818 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.744860 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.744870 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.744886 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.744896 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.838205 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:02 crc kubenswrapper[4867]: E1212 06:50:02.838334 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.838529 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:02 crc kubenswrapper[4867]: E1212 06:50:02.838585 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.838730 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:02 crc kubenswrapper[4867]: E1212 06:50:02.838809 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.847130 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.847179 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.847194 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.847216 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.847247 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.863208 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.863191003 podStartE2EDuration="35.863191003s" podCreationTimestamp="2025-12-12 06:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:02.862778712 +0000 UTC m=+90.434159981" watchObservedRunningTime="2025-12-12 06:50:02.863191003 +0000 UTC m=+90.434572272" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.905343 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.90532437 podStartE2EDuration="1m11.90532437s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:02.888303748 +0000 UTC m=+90.459685017" watchObservedRunningTime="2025-12-12 06:50:02.90532437 +0000 UTC m=+90.476705639" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.916686 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dsjbr" podStartSLOduration=72.916661963 podStartE2EDuration="1m12.916661963s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:02.916606232 +0000 UTC m=+90.487987511" watchObservedRunningTime="2025-12-12 06:50:02.916661963 +0000 UTC m=+90.488043242" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.947155 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.947134886 podStartE2EDuration="1m9.947134886s" podCreationTimestamp="2025-12-12 06:48:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:02.931964167 +0000 UTC m=+90.503345436" watchObservedRunningTime="2025-12-12 06:50:02.947134886 +0000 UTC m=+90.518516155" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.952522 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.952569 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.952583 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.952604 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.952616 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:02Z","lastTransitionTime":"2025-12-12T06:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:02 crc kubenswrapper[4867]: I1212 06:50:02.992568 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podStartSLOduration=72.992548743 podStartE2EDuration="1m12.992548743s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:02.992166372 +0000 UTC m=+90.563547651" watchObservedRunningTime="2025-12-12 06:50:02.992548743 +0000 UTC m=+90.563930012" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.048344 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=69.048326856 podStartE2EDuration="1m9.048326856s" podCreationTimestamp="2025-12-12 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:03.039121651 +0000 UTC m=+90.610502920" watchObservedRunningTime="2025-12-12 06:50:03.048326856 +0000 UTC m=+90.619708125" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.055137 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.055161 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.055171 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.055188 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.055198 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.059754 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-njfkw" podStartSLOduration=73.059744482 podStartE2EDuration="1m13.059744482s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:03.048937473 +0000 UTC m=+90.620318742" watchObservedRunningTime="2025-12-12 06:50:03.059744482 +0000 UTC m=+90.631125751" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.075428 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9dpkr" podStartSLOduration=73.075398805 podStartE2EDuration="1m13.075398805s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:03.073751729 +0000 UTC m=+90.645132998" watchObservedRunningTime="2025-12-12 06:50:03.075398805 +0000 UTC m=+90.646780094" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.100392 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-z8lss" podStartSLOduration=73.100369546 podStartE2EDuration="1m13.100369546s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:03.08970004 +0000 UTC m=+90.661081299" watchObservedRunningTime="2025-12-12 06:50:03.100369546 +0000 UTC m=+90.671750835" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.101198 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vrtlz" podStartSLOduration=72.101192048 podStartE2EDuration="1m12.101192048s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:03.099938384 +0000 UTC m=+90.671319663" watchObservedRunningTime="2025-12-12 06:50:03.101192048 +0000 UTC m=+90.672573337" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.157571 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.157608 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.157616 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.157631 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.157641 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.259057 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.259097 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.259107 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.259122 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.259134 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.361371 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.361420 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.361434 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.361451 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.361464 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.463813 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.463850 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.463858 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.463873 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.463882 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.566589 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.566637 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.566645 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.566659 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.566668 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.668761 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.668822 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.668842 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.668911 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.668948 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.772048 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.772092 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.772107 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.772128 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.772147 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.837829 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:03 crc kubenswrapper[4867]: E1212 06:50:03.838067 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.850120 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.874446 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.874484 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.874493 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.874510 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.874523 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.978030 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.978087 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.978101 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.978123 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:03 crc kubenswrapper[4867]: I1212 06:50:03.978140 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:03Z","lastTransitionTime":"2025-12-12T06:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.081088 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.081154 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.081170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.081204 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.081221 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.184079 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.184170 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.184190 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.184223 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.184280 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.287292 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.287367 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.287386 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.287415 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.287437 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.390092 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.390153 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.390165 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.390185 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.390196 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.493457 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.493541 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.493562 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.493594 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.493617 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.595749 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.595788 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.595798 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.595812 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.595821 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.697380 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.697417 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.697425 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.697438 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.697447 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.799801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.799848 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.799866 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.799883 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.799896 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.838150 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:04 crc kubenswrapper[4867]: E1212 06:50:04.838663 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.838324 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:04 crc kubenswrapper[4867]: E1212 06:50:04.839503 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.838285 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:04 crc kubenswrapper[4867]: E1212 06:50:04.839861 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.901892 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.901974 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.901985 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.902001 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:04 crc kubenswrapper[4867]: I1212 06:50:04.902011 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:04Z","lastTransitionTime":"2025-12-12T06:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.004377 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.004413 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.004424 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.004448 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.004459 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.106738 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.106810 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.106832 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.106859 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.106877 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.209801 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.209877 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.209913 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.209943 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.209964 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.313265 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.313295 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.313303 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.313318 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.313326 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.416799 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.416840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.416849 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.416869 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.416878 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.520274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.520319 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.520332 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.520348 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.520358 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.622781 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.622816 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.622823 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.622838 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.622847 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.725706 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.725840 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.725862 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.725890 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.725908 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.829185 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.829244 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.829258 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.829274 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.829286 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.837720 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:05 crc kubenswrapper[4867]: E1212 06:50:05.837943 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.855566 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.855625 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.855643 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.855667 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.855682 4867 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-12T06:50:05Z","lastTransitionTime":"2025-12-12T06:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.915571 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc"] Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.915956 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.917712 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.917787 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.918795 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.927011 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 12 06:50:05 crc kubenswrapper[4867]: I1212 06:50:05.939041 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.939013105 podStartE2EDuration="2.939013105s" podCreationTimestamp="2025-12-12 06:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:05.93807622 +0000 UTC m=+93.509457519" watchObservedRunningTime="2025-12-12 06:50:05.939013105 +0000 UTC m=+93.510394414" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.060469 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.060515 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.060542 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e9f489b-8429-4543-a613-ca43bfc1eeff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.060556 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6e9f489b-8429-4543-a613-ca43bfc1eeff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.060586 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e9f489b-8429-4543-a613-ca43bfc1eeff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.161963 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162022 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162061 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e9f489b-8429-4543-a613-ca43bfc1eeff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162114 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6e9f489b-8429-4543-a613-ca43bfc1eeff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162190 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e9f489b-8429-4543-a613-ca43bfc1eeff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162480 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.162540 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/6e9f489b-8429-4543-a613-ca43bfc1eeff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.164202 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6e9f489b-8429-4543-a613-ca43bfc1eeff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.167997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e9f489b-8429-4543-a613-ca43bfc1eeff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.190274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e9f489b-8429-4543-a613-ca43bfc1eeff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-zwltc\" (UID: \"6e9f489b-8429-4543-a613-ca43bfc1eeff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.239673 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.358470 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" event={"ID":"6e9f489b-8429-4543-a613-ca43bfc1eeff","Type":"ContainerStarted","Data":"972903fe286e6466a76e82a4ef5c24626131f1c43790ee985a9cfd3c8417a63f"} Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.837742 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.837872 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:06 crc kubenswrapper[4867]: E1212 06:50:06.837877 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:06 crc kubenswrapper[4867]: E1212 06:50:06.838058 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:06 crc kubenswrapper[4867]: I1212 06:50:06.838180 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:06 crc kubenswrapper[4867]: E1212 06:50:06.838353 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:07 crc kubenswrapper[4867]: I1212 06:50:07.364207 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" event={"ID":"6e9f489b-8429-4543-a613-ca43bfc1eeff","Type":"ContainerStarted","Data":"d0bcb5c8336ff47eb1b4f30c6c3fcbefcf035732f6c9dc1476f19640567842d9"} Dec 12 06:50:07 crc kubenswrapper[4867]: I1212 06:50:07.385265 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-zwltc" podStartSLOduration=77.385194199 podStartE2EDuration="1m17.385194199s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:07.383186614 +0000 UTC m=+94.954567903" watchObservedRunningTime="2025-12-12 06:50:07.385194199 +0000 UTC m=+94.956575518" Dec 12 06:50:07 crc kubenswrapper[4867]: I1212 06:50:07.837760 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:07 crc kubenswrapper[4867]: E1212 06:50:07.838462 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:08 crc kubenswrapper[4867]: I1212 06:50:08.586362 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:08 crc kubenswrapper[4867]: E1212 06:50:08.586531 4867 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:50:08 crc kubenswrapper[4867]: E1212 06:50:08.586594 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs podName:c4f347fa-32d5-44e0-bc90-6d774fa43ce0 nodeName:}" failed. No retries permitted until 2025-12-12 06:51:12.586576659 +0000 UTC m=+160.157957928 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs") pod "network-metrics-daemon-hj2lv" (UID: "c4f347fa-32d5-44e0-bc90-6d774fa43ce0") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 12 06:50:08 crc kubenswrapper[4867]: I1212 06:50:08.837535 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:08 crc kubenswrapper[4867]: E1212 06:50:08.837903 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:08 crc kubenswrapper[4867]: I1212 06:50:08.837557 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:08 crc kubenswrapper[4867]: E1212 06:50:08.838139 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:08 crc kubenswrapper[4867]: I1212 06:50:08.837535 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:08 crc kubenswrapper[4867]: E1212 06:50:08.838380 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:09 crc kubenswrapper[4867]: I1212 06:50:09.837091 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:09 crc kubenswrapper[4867]: E1212 06:50:09.837256 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:10 crc kubenswrapper[4867]: I1212 06:50:10.837906 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:10 crc kubenswrapper[4867]: I1212 06:50:10.837974 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:10 crc kubenswrapper[4867]: I1212 06:50:10.838013 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:10 crc kubenswrapper[4867]: E1212 06:50:10.838141 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:10 crc kubenswrapper[4867]: E1212 06:50:10.838326 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:10 crc kubenswrapper[4867]: E1212 06:50:10.838605 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:11 crc kubenswrapper[4867]: I1212 06:50:11.837058 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:11 crc kubenswrapper[4867]: E1212 06:50:11.837221 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:12 crc kubenswrapper[4867]: I1212 06:50:12.837929 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:12 crc kubenswrapper[4867]: I1212 06:50:12.838364 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:12 crc kubenswrapper[4867]: I1212 06:50:12.838442 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:12 crc kubenswrapper[4867]: E1212 06:50:12.839757 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:12 crc kubenswrapper[4867]: E1212 06:50:12.839817 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:12 crc kubenswrapper[4867]: E1212 06:50:12.839871 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:13 crc kubenswrapper[4867]: I1212 06:50:13.837172 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:13 crc kubenswrapper[4867]: E1212 06:50:13.837466 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:13 crc kubenswrapper[4867]: I1212 06:50:13.838545 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 06:50:13 crc kubenswrapper[4867]: E1212 06:50:13.838841 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbsqz_openshift-ovn-kubernetes(5f519f9a-7a2d-4193-8c21-ae7840036dae)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" Dec 12 06:50:14 crc kubenswrapper[4867]: I1212 06:50:14.837588 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:14 crc kubenswrapper[4867]: I1212 06:50:14.837652 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:14 crc kubenswrapper[4867]: I1212 06:50:14.837664 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:14 crc kubenswrapper[4867]: E1212 06:50:14.837743 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:14 crc kubenswrapper[4867]: E1212 06:50:14.837896 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:14 crc kubenswrapper[4867]: E1212 06:50:14.837975 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:15 crc kubenswrapper[4867]: I1212 06:50:15.837897 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:15 crc kubenswrapper[4867]: E1212 06:50:15.838042 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:16 crc kubenswrapper[4867]: I1212 06:50:16.837500 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:16 crc kubenswrapper[4867]: E1212 06:50:16.837697 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:16 crc kubenswrapper[4867]: I1212 06:50:16.837979 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:16 crc kubenswrapper[4867]: E1212 06:50:16.838088 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:16 crc kubenswrapper[4867]: I1212 06:50:16.838462 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:16 crc kubenswrapper[4867]: E1212 06:50:16.838562 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:17 crc kubenswrapper[4867]: I1212 06:50:17.837976 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:17 crc kubenswrapper[4867]: E1212 06:50:17.838134 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:18 crc kubenswrapper[4867]: I1212 06:50:18.838155 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:18 crc kubenswrapper[4867]: I1212 06:50:18.838340 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:18 crc kubenswrapper[4867]: I1212 06:50:18.838155 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:18 crc kubenswrapper[4867]: E1212 06:50:18.838550 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:18 crc kubenswrapper[4867]: E1212 06:50:18.838722 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:18 crc kubenswrapper[4867]: E1212 06:50:18.838864 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:19 crc kubenswrapper[4867]: I1212 06:50:19.837418 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:19 crc kubenswrapper[4867]: E1212 06:50:19.837844 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:20 crc kubenswrapper[4867]: I1212 06:50:20.838387 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:20 crc kubenswrapper[4867]: I1212 06:50:20.838446 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:20 crc kubenswrapper[4867]: E1212 06:50:20.838713 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:20 crc kubenswrapper[4867]: E1212 06:50:20.838865 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:20 crc kubenswrapper[4867]: I1212 06:50:20.839212 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:20 crc kubenswrapper[4867]: E1212 06:50:20.839407 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:21 crc kubenswrapper[4867]: I1212 06:50:21.837537 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:21 crc kubenswrapper[4867]: E1212 06:50:21.837658 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:22 crc kubenswrapper[4867]: I1212 06:50:22.838522 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:22 crc kubenswrapper[4867]: I1212 06:50:22.838776 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:22 crc kubenswrapper[4867]: I1212 06:50:22.838964 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:22 crc kubenswrapper[4867]: E1212 06:50:22.843217 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:22 crc kubenswrapper[4867]: E1212 06:50:22.843436 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:22 crc kubenswrapper[4867]: E1212 06:50:22.843926 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.421728 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/1.log" Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.422515 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/0.log" Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.422572 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1918c53-1a6d-4317-a320-2e301b5a9db7" containerID="9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9" exitCode=1 Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.422608 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerDied","Data":"9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9"} Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.422675 4867 scope.go:117] "RemoveContainer" containerID="e79e2a7b16596d59444d30e642de47443997410893cca2a0f02d3ced70f0bdfd" Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.423154 4867 scope.go:117] "RemoveContainer" containerID="9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9" Dec 12 06:50:23 crc kubenswrapper[4867]: E1212 06:50:23.423353 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-z8lss_openshift-multus(f1918c53-1a6d-4317-a320-2e301b5a9db7)\"" pod="openshift-multus/multus-z8lss" podUID="f1918c53-1a6d-4317-a320-2e301b5a9db7" Dec 12 06:50:23 crc kubenswrapper[4867]: I1212 06:50:23.837088 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:23 crc kubenswrapper[4867]: E1212 06:50:23.837220 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:24 crc kubenswrapper[4867]: I1212 06:50:24.426405 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/1.log" Dec 12 06:50:24 crc kubenswrapper[4867]: I1212 06:50:24.837976 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:24 crc kubenswrapper[4867]: I1212 06:50:24.838060 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:24 crc kubenswrapper[4867]: I1212 06:50:24.838022 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:24 crc kubenswrapper[4867]: E1212 06:50:24.838172 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:24 crc kubenswrapper[4867]: E1212 06:50:24.839316 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:24 crc kubenswrapper[4867]: E1212 06:50:24.839592 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:25 crc kubenswrapper[4867]: I1212 06:50:25.837424 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:25 crc kubenswrapper[4867]: E1212 06:50:25.837653 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:26 crc kubenswrapper[4867]: I1212 06:50:26.837718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:26 crc kubenswrapper[4867]: I1212 06:50:26.837876 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:26 crc kubenswrapper[4867]: E1212 06:50:26.837972 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:26 crc kubenswrapper[4867]: I1212 06:50:26.837753 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:26 crc kubenswrapper[4867]: E1212 06:50:26.838144 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:26 crc kubenswrapper[4867]: E1212 06:50:26.838321 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:27 crc kubenswrapper[4867]: I1212 06:50:27.837888 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:27 crc kubenswrapper[4867]: E1212 06:50:27.838059 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:28 crc kubenswrapper[4867]: I1212 06:50:28.837585 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:28 crc kubenswrapper[4867]: I1212 06:50:28.837679 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:28 crc kubenswrapper[4867]: E1212 06:50:28.837804 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:28 crc kubenswrapper[4867]: E1212 06:50:28.837855 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:28 crc kubenswrapper[4867]: I1212 06:50:28.838152 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:28 crc kubenswrapper[4867]: E1212 06:50:28.838274 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:28 crc kubenswrapper[4867]: I1212 06:50:28.838554 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.444893 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/3.log" Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.447271 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerStarted","Data":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.447972 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.476021 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podStartSLOduration=99.476003543 podStartE2EDuration="1m39.476003543s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:29.475401828 +0000 UTC m=+117.046783117" watchObservedRunningTime="2025-12-12 06:50:29.476003543 +0000 UTC m=+117.047384812" Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.711704 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hj2lv"] Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.711937 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:29 crc kubenswrapper[4867]: E1212 06:50:29.712145 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:29 crc kubenswrapper[4867]: I1212 06:50:29.837852 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:29 crc kubenswrapper[4867]: E1212 06:50:29.837971 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:30 crc kubenswrapper[4867]: I1212 06:50:30.837942 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:30 crc kubenswrapper[4867]: E1212 06:50:30.838407 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:30 crc kubenswrapper[4867]: I1212 06:50:30.837989 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:30 crc kubenswrapper[4867]: E1212 06:50:30.838488 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:31 crc kubenswrapper[4867]: I1212 06:50:31.837092 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:31 crc kubenswrapper[4867]: I1212 06:50:31.837271 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:31 crc kubenswrapper[4867]: E1212 06:50:31.837276 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:31 crc kubenswrapper[4867]: E1212 06:50:31.837503 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:32 crc kubenswrapper[4867]: E1212 06:50:32.814285 4867 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 12 06:50:32 crc kubenswrapper[4867]: I1212 06:50:32.837563 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:32 crc kubenswrapper[4867]: I1212 06:50:32.837596 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:32 crc kubenswrapper[4867]: E1212 06:50:32.839029 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:32 crc kubenswrapper[4867]: E1212 06:50:32.839201 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:32 crc kubenswrapper[4867]: E1212 06:50:32.932270 4867 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 12 06:50:33 crc kubenswrapper[4867]: I1212 06:50:33.837535 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:33 crc kubenswrapper[4867]: I1212 06:50:33.837535 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:33 crc kubenswrapper[4867]: E1212 06:50:33.837738 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:33 crc kubenswrapper[4867]: E1212 06:50:33.837845 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:34 crc kubenswrapper[4867]: I1212 06:50:34.837679 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:34 crc kubenswrapper[4867]: I1212 06:50:34.837848 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:34 crc kubenswrapper[4867]: E1212 06:50:34.837857 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:34 crc kubenswrapper[4867]: E1212 06:50:34.838110 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:34 crc kubenswrapper[4867]: I1212 06:50:34.838684 4867 scope.go:117] "RemoveContainer" containerID="9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9" Dec 12 06:50:35 crc kubenswrapper[4867]: I1212 06:50:35.470286 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/1.log" Dec 12 06:50:35 crc kubenswrapper[4867]: I1212 06:50:35.470356 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerStarted","Data":"fb4548196b76adb7c889aee7392b15bcfde6b7d6613c5b208454e042140f1889"} Dec 12 06:50:35 crc kubenswrapper[4867]: I1212 06:50:35.837539 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:35 crc kubenswrapper[4867]: I1212 06:50:35.837598 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:35 crc kubenswrapper[4867]: E1212 06:50:35.837966 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:35 crc kubenswrapper[4867]: E1212 06:50:35.838140 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:36 crc kubenswrapper[4867]: I1212 06:50:36.837551 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:36 crc kubenswrapper[4867]: I1212 06:50:36.837903 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:36 crc kubenswrapper[4867]: E1212 06:50:36.838295 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 12 06:50:36 crc kubenswrapper[4867]: E1212 06:50:36.838502 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 12 06:50:37 crc kubenswrapper[4867]: I1212 06:50:37.837822 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:37 crc kubenswrapper[4867]: I1212 06:50:37.837896 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:37 crc kubenswrapper[4867]: E1212 06:50:37.837986 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hj2lv" podUID="c4f347fa-32d5-44e0-bc90-6d774fa43ce0" Dec 12 06:50:37 crc kubenswrapper[4867]: E1212 06:50:37.838114 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 12 06:50:38 crc kubenswrapper[4867]: I1212 06:50:38.837756 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:38 crc kubenswrapper[4867]: I1212 06:50:38.837821 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:38 crc kubenswrapper[4867]: I1212 06:50:38.841842 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 12 06:50:38 crc kubenswrapper[4867]: I1212 06:50:38.841931 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.838113 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.838219 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.841593 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.841702 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.841587 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 12 06:50:39 crc kubenswrapper[4867]: I1212 06:50:39.841902 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.577800 4867 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.606804 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.607305 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.609259 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.609523 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.609650 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.609755 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.620036 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-nmwfj"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.620380 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.624553 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.624785 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.627049 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.628527 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.629101 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.629322 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.629858 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.630529 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-52qk7"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.630939 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.633288 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjj4n"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.633663 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.633781 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.634384 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.635275 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.635624 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.636164 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.636176 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.636478 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.636886 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.637274 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.637501 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.637531 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.638501 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.639035 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.639672 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.639960 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.648960 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.652981 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.652994 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653324 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653532 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653651 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653709 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653851 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653870 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653948 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.653965 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.654029 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.654042 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.654113 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.655290 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.655837 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.656070 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.660953 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.661442 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.662194 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.662375 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.662587 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.662765 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663274 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lrj5h"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663434 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663637 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663713 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663750 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663803 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.663974 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664085 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664087 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5r5vl"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664257 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664273 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664365 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664444 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664501 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664555 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664644 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664667 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664686 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.664838 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.665040 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.665135 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.665154 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.665164 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.665565 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.666323 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.680200 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.680611 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.680845 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.688664 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.688822 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.689129 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.689774 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.690285 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.690454 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs52"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.690529 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.690982 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.690997 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.691116 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.691349 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.691384 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.691504 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.691614 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692072 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692262 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692584 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692680 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692913 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.692997 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.693034 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lk4nq"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.693420 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.693708 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.694589 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfk5k"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.694937 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.695530 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.696898 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.697034 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.697438 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.697574 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.698402 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.698734 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699033 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699095 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699241 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699318 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699347 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699411 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699532 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699615 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699638 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699664 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699710 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.699729 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.700087 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.701084 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.701550 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6gl7z"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.701985 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.702004 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.702207 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.702321 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.702906 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.703157 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s2k8\" (UniqueName: \"kubernetes.io/projected/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-kube-api-access-2s2k8\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.703255 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.705438 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.706153 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.706512 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7zlz5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.706949 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.709937 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.713497 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.713645 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.713779 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.714105 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-592qx"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.714453 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.714568 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.714746 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.714875 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.715059 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.715598 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.717534 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.718355 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.718718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.718849 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.724831 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.725636 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.726036 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.734030 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.736839 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c7vxj"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.739172 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.740011 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.750966 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.762389 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.764253 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.764614 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.764756 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.765027 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.765385 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.765728 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.765750 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.766284 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.767121 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.767391 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.768789 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.771276 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.771894 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.772289 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.772428 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.772721 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.785894 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.789500 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.791962 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.798973 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.800294 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjj4n"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.802467 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nmwfj"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.802707 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804035 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804067 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-images\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804088 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804106 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjgjz\" (UniqueName: \"kubernetes.io/projected/818edab4-a66c-4739-bead-9d42292e468a-kube-api-access-cjgjz\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804125 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-serving-cert\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804139 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804156 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804173 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804188 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804204 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g2kq\" (UniqueName: \"kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804219 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-serving-cert\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804358 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804375 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ltzh\" (UniqueName: \"kubernetes.io/projected/2ab99367-5422-43bd-97d9-550e989d7542-kube-api-access-5ltzh\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804390 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804408 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9rzz\" (UniqueName: \"kubernetes.io/projected/f1ad2733-54b4-4365-9651-c6dbb04d7493-kube-api-access-m9rzz\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804431 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-config\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97nkc\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-kube-api-access-97nkc\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804472 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/818edab4-a66c-4739-bead-9d42292e468a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804523 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c1e11f9-9309-4412-bda9-8d39d3192749-trusted-ca\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804539 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtvsc\" (UniqueName: \"kubernetes.io/projected/45a95b26-1418-4107-bd27-0f67e3e6b9db-kube-api-access-rtvsc\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804553 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804568 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-service-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804584 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804601 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/865aeb86-6c89-4b2b-88c6-4d9637442caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804616 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804630 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e33a71af-948c-4fe9-a374-d68bbb4fc862-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804679 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv77x\" (UniqueName: \"kubernetes.io/projected/b9ce7d0e-4827-492c-a71e-53299557a2b5-kube-api-access-qv77x\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804698 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1ad2733-54b4-4365-9651-c6dbb04d7493-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804715 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804730 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804752 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804785 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-machine-approver-tls\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804802 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9wrx\" (UniqueName: \"kubernetes.io/projected/865aeb86-6c89-4b2b-88c6-4d9637442caa-kube-api-access-r9wrx\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804848 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-client\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804865 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-client\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804880 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-config\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804896 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804914 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804928 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj8b\" (UniqueName: \"kubernetes.io/projected/7b0920c8-ab26-4b72-b22b-c4123498a461-kube-api-access-qsj8b\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804945 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn59v\" (UniqueName: \"kubernetes.io/projected/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-kube-api-access-rn59v\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804961 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804974 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a2028ce-f6e9-414f-b071-301f03fa6337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805024 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44957455-bf87-4493-bee7-0304b2165061-serving-cert\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805038 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc67p\" (UniqueName: \"kubernetes.io/projected/44957455-bf87-4493-bee7-0304b2165061-kube-api-access-lc67p\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805052 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdqr5\" (UniqueName: \"kubernetes.io/projected/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-kube-api-access-sdqr5\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805067 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae7c8f-1ae7-443f-a049-7eca845b9387-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805086 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ae7c8f-1ae7-443f-a049-7eca845b9387-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805108 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805125 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805139 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805189 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805206 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-service-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805233 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhkd2\" (UniqueName: \"kubernetes.io/projected/bd04d7ab-8933-4815-97ec-7be8b2e643ce-kube-api-access-bhkd2\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805250 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805266 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7wk\" (UniqueName: \"kubernetes.io/projected/2a2028ce-f6e9-414f-b071-301f03fa6337-kube-api-access-tf7wk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805283 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5265271c-7232-4ce2-b0aa-fbfe5330e267-serving-cert\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805299 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805345 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd04d7ab-8933-4815-97ec-7be8b2e643ce-service-ca-bundle\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805363 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805378 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-auth-proxy-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805395 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805410 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e33a71af-948c-4fe9-a374-d68bbb4fc862-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818edab4-a66c-4739-bead-9d42292e468a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805443 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805459 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2028ce-f6e9-414f-b071-301f03fa6337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805473 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-stats-auth\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805521 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-images\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805543 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93f987df-943c-45a9-aaf8-a0430358ba7f-proxy-tls\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-srv-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805593 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805616 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805632 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-key\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805645 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805696 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805711 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ab99367-5422-43bd-97d9-550e989d7542-metrics-tls\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805726 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-trusted-ca\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805742 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ce7d0e-4827-492c-a71e-53299557a2b5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805756 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805775 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbmt\" (UniqueName: \"kubernetes.io/projected/9aab65fa-d90b-489e-8443-07cac98cb8ee-kube-api-access-ggbmt\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805788 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805810 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44957455-bf87-4493-bee7-0304b2165061-config\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805851 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28mcc\" (UniqueName: \"kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805866 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-cabundle\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805882 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jdj\" (UniqueName: \"kubernetes.io/projected/db9be553-1a67-4dc9-874b-7b79d332fcce-kube-api-access-79jdj\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805905 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9c1e11f9-9309-4412-bda9-8d39d3192749-metrics-tls\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805918 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qn79\" (UniqueName: \"kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805935 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcbzb\" (UniqueName: \"kubernetes.io/projected/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-kube-api-access-gcbzb\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805950 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/865aeb86-6c89-4b2b-88c6-4d9637442caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.805965 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ae7c8f-1ae7-443f-a049-7eca845b9387-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806002 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlth\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-kube-api-access-gdlth\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806019 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-policies\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806034 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-encryption-config\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806049 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806064 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806081 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-default-certificate\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806102 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s2k8\" (UniqueName: \"kubernetes.io/projected/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-kube-api-access-2s2k8\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806122 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-dir\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806167 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806192 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806207 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-profile-collector-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806244 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-config\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806263 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-config\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806277 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjw9\" (UniqueName: \"kubernetes.io/projected/5265271c-7232-4ce2-b0aa-fbfe5330e267-kube-api-access-zxjw9\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806292 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-metrics-certs\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806330 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806348 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45a95b26-1418-4107-bd27-0f67e3e6b9db-serving-cert\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806362 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjcdp\" (UniqueName: \"kubernetes.io/projected/93f987df-943c-45a9-aaf8-a0430358ba7f-kube-api-access-mjcdp\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806379 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxft9\" (UniqueName: \"kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806395 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lcxs\" (UniqueName: \"kubernetes.io/projected/a648d31a-4689-434c-b89b-c0d96065ce10-kube-api-access-7lcxs\") pod \"downloads-7954f5f757-nmwfj\" (UID: \"a648d31a-4689-434c-b89b-c0d96065ce10\") " pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806409 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-config\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.804695 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.806614 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.810564 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ht7nj"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.811845 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.813632 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.816094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.818830 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nfwlh"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.819649 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.820176 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.822932 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lk4nq"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.828840 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.834057 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.835150 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-52qk7"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.836642 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.845849 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.845893 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.845915 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.845928 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6gl7z"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.847571 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.849673 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.851179 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.851368 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5r5vl"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.854718 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7zlz5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.856183 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.858117 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-592qx"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.860254 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.861697 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.863025 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.864477 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.865862 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.867202 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs52"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.868591 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.869591 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.869989 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.871545 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.873822 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfk5k"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.875941 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c7vxj"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.877303 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-fzxx5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.878028 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.878559 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8lml"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.879817 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.880057 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nfwlh"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.883051 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.884469 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fzxx5"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.885914 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8lml"] Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.889600 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.907956 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv77x\" (UniqueName: \"kubernetes.io/projected/b9ce7d0e-4827-492c-a71e-53299557a2b5-kube-api-access-qv77x\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908055 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1ad2733-54b4-4365-9651-c6dbb04d7493-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908148 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908174 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908191 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908210 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908251 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-machine-approver-tls\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908270 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908288 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-client\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908303 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-config\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908363 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj8b\" (UniqueName: \"kubernetes.io/projected/7b0920c8-ab26-4b72-b22b-c4123498a461-kube-api-access-qsj8b\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908401 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44957455-bf87-4493-bee7-0304b2165061-serving-cert\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908421 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc67p\" (UniqueName: \"kubernetes.io/projected/44957455-bf87-4493-bee7-0304b2165061-kube-api-access-lc67p\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908437 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdqr5\" (UniqueName: \"kubernetes.io/projected/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-kube-api-access-sdqr5\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908454 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae7c8f-1ae7-443f-a049-7eca845b9387-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ae7c8f-1ae7-443f-a049-7eca845b9387-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908490 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908523 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8091880-bdd4-4a8d-8078-dec177dd01bd-proxy-tls\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908543 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhkd2\" (UniqueName: \"kubernetes.io/projected/bd04d7ab-8933-4815-97ec-7be8b2e643ce-kube-api-access-bhkd2\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908559 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908576 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908595 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908616 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd04d7ab-8933-4815-97ec-7be8b2e643ce-service-ca-bundle\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908634 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908652 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-auth-proxy-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908684 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908701 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e33a71af-948c-4fe9-a374-d68bbb4fc862-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908718 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908734 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-srv-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908750 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-stats-auth\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908768 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908786 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908802 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908819 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908834 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ab99367-5422-43bd-97d9-550e989d7542-metrics-tls\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908848 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-trusted-ca\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908864 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ce7d0e-4827-492c-a71e-53299557a2b5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908879 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbmt\" (UniqueName: \"kubernetes.io/projected/9aab65fa-d90b-489e-8443-07cac98cb8ee-kube-api-access-ggbmt\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908894 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908910 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8091880-bdd4-4a8d-8078-dec177dd01bd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908932 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-cabundle\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908954 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jdj\" (UniqueName: \"kubernetes.io/projected/db9be553-1a67-4dc9-874b-7b79d332fcce-kube-api-access-79jdj\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908977 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcbzb\" (UniqueName: \"kubernetes.io/projected/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-kube-api-access-gcbzb\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.908997 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/865aeb86-6c89-4b2b-88c6-4d9637442caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909018 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-policies\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909116 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909143 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-config\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909165 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-dir\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909186 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909205 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909207 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909245 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjw9\" (UniqueName: \"kubernetes.io/projected/5265271c-7232-4ce2-b0aa-fbfe5330e267-kube-api-access-zxjw9\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909443 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-metrics-certs\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45a95b26-1418-4107-bd27-0f67e3e6b9db-serving-cert\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909505 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxft9\" (UniqueName: \"kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909537 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lcxs\" (UniqueName: \"kubernetes.io/projected/a648d31a-4689-434c-b89b-c0d96065ce10-kube-api-access-7lcxs\") pod \"downloads-7954f5f757-nmwfj\" (UID: \"a648d31a-4689-434c-b89b-c0d96065ce10\") " pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909564 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-images\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909591 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909616 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909641 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g2kq\" (UniqueName: \"kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909667 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-serving-cert\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909696 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909726 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ltzh\" (UniqueName: \"kubernetes.io/projected/2ab99367-5422-43bd-97d9-550e989d7542-kube-api-access-5ltzh\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909751 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-config\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909778 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97nkc\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-kube-api-access-97nkc\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909805 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c1e11f9-9309-4412-bda9-8d39d3192749-trusted-ca\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.909829 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtvsc\" (UniqueName: \"kubernetes.io/projected/45a95b26-1418-4107-bd27-0f67e3e6b9db-kube-api-access-rtvsc\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.910540 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ae7c8f-1ae7-443f-a049-7eca845b9387-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.910865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bd04d7ab-8933-4815-97ec-7be8b2e643ce-service-ca-bundle\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.910913 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-config\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911400 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911433 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911504 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/865aeb86-6c89-4b2b-88c6-4d9637442caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911532 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e33a71af-948c-4fe9-a374-d68bbb4fc862-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911568 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9wrx\" (UniqueName: \"kubernetes.io/projected/865aeb86-6c89-4b2b-88c6-4d9637442caa-kube-api-access-r9wrx\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911619 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-client\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911642 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911667 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn59v\" (UniqueName: \"kubernetes.io/projected/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-kube-api-access-rn59v\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911693 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911716 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a2028ce-f6e9-414f-b071-301f03fa6337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911760 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911803 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-service-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911837 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7wk\" (UniqueName: \"kubernetes.io/projected/2a2028ce-f6e9-414f-b071-301f03fa6337-kube-api-access-tf7wk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911865 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5265271c-7232-4ce2-b0aa-fbfe5330e267-serving-cert\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911893 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911921 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818edab4-a66c-4739-bead-9d42292e468a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911948 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.911986 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2028ce-f6e9-414f-b071-301f03fa6337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912011 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-images\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912035 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93f987df-943c-45a9-aaf8-a0430358ba7f-proxy-tls\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912061 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-key\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912087 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912127 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44957455-bf87-4493-bee7-0304b2165061-config\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912165 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28mcc\" (UniqueName: \"kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912192 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9c1e11f9-9309-4412-bda9-8d39d3192749-metrics-tls\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912221 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qn79\" (UniqueName: \"kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912270 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlth\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-kube-api-access-gdlth\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912299 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ae7c8f-1ae7-443f-a049-7eca845b9387-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912326 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-default-certificate\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912352 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-encryption-config\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912378 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912411 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-profile-collector-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912463 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-config\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912489 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjcdp\" (UniqueName: \"kubernetes.io/projected/93f987df-943c-45a9-aaf8-a0430358ba7f-kube-api-access-mjcdp\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912517 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-config\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912542 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmnl\" (UniqueName: \"kubernetes.io/projected/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-kube-api-access-gnmnl\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912573 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjgjz\" (UniqueName: \"kubernetes.io/projected/818edab4-a66c-4739-bead-9d42292e468a-kube-api-access-cjgjz\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912628 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912655 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sd9l\" (UniqueName: \"kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912687 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-serving-cert\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912714 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912739 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912766 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912794 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912821 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9rzz\" (UniqueName: \"kubernetes.io/projected/f1ad2733-54b4-4365-9651-c6dbb04d7493-kube-api-access-m9rzz\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912847 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-config\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912871 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw5wb\" (UniqueName: \"kubernetes.io/projected/a8091880-bdd4-4a8d-8078-dec177dd01bd-kube-api-access-zw5wb\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912877 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-auth-proxy-config\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912901 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/818edab4-a66c-4739-bead-9d42292e468a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912928 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-service-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912955 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.913558 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/44957455-bf87-4493-bee7-0304b2165061-serving-cert\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.913709 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.913920 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.914075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-machine-approver-tls\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.914295 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.914626 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a2028ce-f6e9-414f-b071-301f03fa6337-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915111 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915113 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915418 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/865aeb86-6c89-4b2b-88c6-4d9637442caa-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915735 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45a95b26-1418-4107-bd27-0f67e3e6b9db-serving-cert\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.915993 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e33a71af-948c-4fe9-a374-d68bbb4fc862-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.916014 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-stats-auth\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.916146 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/45a95b26-1418-4107-bd27-0f67e3e6b9db-trusted-ca\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.916400 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.916798 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.917056 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-cabundle\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.917090 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.917135 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/93f987df-943c-45a9-aaf8-a0430358ba7f-images\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.917368 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-config\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.912269 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.918513 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/818edab4-a66c-4739-bead-9d42292e468a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919044 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-config\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919295 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44957455-bf87-4493-bee7-0304b2165061-config\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919348 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919411 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-dir\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919610 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-serving-cert\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.919749 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920089 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e33a71af-948c-4fe9-a374-d68bbb4fc862-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920171 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-service-ca\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.917221 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920620 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920679 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ab99367-5422-43bd-97d9-550e989d7542-metrics-tls\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.920718 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-metrics-certs\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.921537 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.922760 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ae7c8f-1ae7-443f-a049-7eca845b9387-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.922780 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/93f987df-943c-45a9-aaf8-a0430358ba7f-proxy-tls\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.922928 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.922985 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/865aeb86-6c89-4b2b-88c6-4d9637442caa-serving-cert\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.923048 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.923484 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.923788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.924327 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7b0920c8-ab26-4b72-b22b-c4123498a461-etcd-client\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.924539 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.924807 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a2028ce-f6e9-414f-b071-301f03fa6337-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.925284 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-profile-collector-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.925481 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/bd04d7ab-8933-4815-97ec-7be8b2e643ce-default-certificate\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.925698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.925711 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/818edab4-a66c-4739-bead-9d42292e468a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.925865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-srv-cert\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.926334 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.926406 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.926795 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f1ad2733-54b4-4365-9651-c6dbb04d7493-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.928530 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.928643 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db9be553-1a67-4dc9-874b-7b79d332fcce-signing-key\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.929589 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.931600 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.933497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.950208 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.969415 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.989956 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 12 06:50:46 crc kubenswrapper[4867]: I1212 06:50:46.997294 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-encryption-config\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.009754 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.013873 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014002 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014090 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmnl\" (UniqueName: \"kubernetes.io/projected/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-kube-api-access-gnmnl\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014112 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sd9l\" (UniqueName: \"kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014167 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-config\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014198 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw5wb\" (UniqueName: \"kubernetes.io/projected/a8091880-bdd4-4a8d-8078-dec177dd01bd-kube-api-access-zw5wb\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014258 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014276 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014320 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8091880-bdd4-4a8d-8078-dec177dd01bd-proxy-tls\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014347 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.014431 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8091880-bdd4-4a8d-8078-dec177dd01bd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.015613 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8091880-bdd4-4a8d-8078-dec177dd01bd-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.029439 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.040781 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-client\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.051795 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.069205 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.079689 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.089112 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.109756 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.119183 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aab65fa-d90b-489e-8443-07cac98cb8ee-serving-cert\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.129549 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.136852 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-audit-policies\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.149235 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.152371 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9aab65fa-d90b-489e-8443-07cac98cb8ee-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.173893 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.176727 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c1e11f9-9309-4412-bda9-8d39d3192749-trusted-ca\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.188802 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.209540 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.222384 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9c1e11f9-9309-4412-bda9-8d39d3192749-metrics-tls\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.229406 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.249552 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.269417 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.290025 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.309451 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.319563 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ce7d0e-4827-492c-a71e-53299557a2b5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.329015 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.337328 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-config\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.350275 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.352788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b9ce7d0e-4827-492c-a71e-53299557a2b5-images\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.374325 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.390679 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.396966 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-service-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.410420 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.423218 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5265271c-7232-4ce2-b0aa-fbfe5330e267-serving-cert\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.430622 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.450256 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.469535 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.489992 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.499299 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-config\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.517422 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.518340 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5265271c-7232-4ce2-b0aa-fbfe5330e267-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.528873 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.533701 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.550063 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.570786 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.589792 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.610746 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.617275 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.630037 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.650070 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.670706 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.678428 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.689576 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.695063 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-config\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.710364 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.728669 4867 request.go:700] Waited for 1.009536886s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmcc-proxy-tls&limit=500&resourceVersion=0 Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.730896 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.738344 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8091880-bdd4-4a8d-8078-dec177dd01bd-proxy-tls\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.749649 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.770347 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.789485 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.798658 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.810125 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.818991 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.831387 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.835959 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.850735 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.870315 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.890821 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.911084 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.929850 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.949304 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.968772 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 12 06:50:47 crc kubenswrapper[4867]: I1212 06:50:47.990479 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.010584 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 12 06:50:48 crc kubenswrapper[4867]: E1212 06:50:48.014649 4867 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 12 06:50:48 crc kubenswrapper[4867]: E1212 06:50:48.014715 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert podName:4f7a80f9-c555-4325-b747-1a5a6e39e6ce nodeName:}" failed. No retries permitted until 2025-12-12 06:50:48.514698541 +0000 UTC m=+136.086079810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert") pod "ingress-canary-nfwlh" (UID: "4f7a80f9-c555-4325-b747-1a5a6e39e6ce") : failed to sync secret cache: timed out waiting for the condition Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.028956 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.049509 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.069397 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.089969 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.116607 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.129336 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.149734 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.168815 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.189832 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.210303 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.235378 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.249781 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.290469 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.311137 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.329845 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.349562 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.368871 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.390095 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.409515 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.450013 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.454139 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s2k8\" (UniqueName: \"kubernetes.io/projected/eecb4b3a-68a9-4e82-8d06-2f06fff6d104-kube-api-access-2s2k8\") pod \"cluster-samples-operator-665b6dd947-lqkgz\" (UID: \"eecb4b3a-68a9-4e82-8d06-2f06fff6d104\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.470868 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.489858 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.509026 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.529390 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.536764 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.542085 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-cert\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.551580 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.569208 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.609423 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.629542 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.649390 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.669283 4867 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.690565 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.709800 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.731210 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.748426 4867 request.go:700] Waited for 1.838997917s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.753222 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv77x\" (UniqueName: \"kubernetes.io/projected/b9ce7d0e-4827-492c-a71e-53299557a2b5-kube-api-access-qv77x\") pod \"machine-api-operator-5694c8668f-7zlz5\" (UID: \"b9ce7d0e-4827-492c-a71e-53299557a2b5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.773392 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjw9\" (UniqueName: \"kubernetes.io/projected/5265271c-7232-4ce2-b0aa-fbfe5330e267-kube-api-access-zxjw9\") pod \"authentication-operator-69f744f599-592qx\" (UID: \"5265271c-7232-4ce2-b0aa-fbfe5330e267\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.775515 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.783701 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.795743 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc67p\" (UniqueName: \"kubernetes.io/projected/44957455-bf87-4493-bee7-0304b2165061-kube-api-access-lc67p\") pod \"service-ca-operator-777779d784-nvs52\" (UID: \"44957455-bf87-4493-bee7-0304b2165061\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.812539 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdqr5\" (UniqueName: \"kubernetes.io/projected/3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a-kube-api-access-sdqr5\") pod \"machine-approver-56656f9798-ppf4t\" (UID: \"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.837324 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e3ae7c8f-1ae7-443f-a049-7eca845b9387-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fjccs\" (UID: \"e3ae7c8f-1ae7-443f-a049-7eca845b9387\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.841436 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.846806 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj8b\" (UniqueName: \"kubernetes.io/projected/7b0920c8-ab26-4b72-b22b-c4123498a461-kube-api-access-qsj8b\") pod \"etcd-operator-b45778765-52qk7\" (UID: \"7b0920c8-ab26-4b72-b22b-c4123498a461\") " pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.861934 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.869110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhkd2\" (UniqueName: \"kubernetes.io/projected/bd04d7ab-8933-4815-97ec-7be8b2e643ce-kube-api-access-bhkd2\") pod \"router-default-5444994796-lrj5h\" (UID: \"bd04d7ab-8933-4815-97ec-7be8b2e643ce\") " pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.889262 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g2kq\" (UniqueName: \"kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq\") pod \"oauth-openshift-558db77b4-bfk5k\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.907895 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lcxs\" (UniqueName: \"kubernetes.io/projected/a648d31a-4689-434c-b89b-c0d96065ce10-kube-api-access-7lcxs\") pod \"downloads-7954f5f757-nmwfj\" (UID: \"a648d31a-4689-434c-b89b-c0d96065ce10\") " pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.924849 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcbzb\" (UniqueName: \"kubernetes.io/projected/bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61-kube-api-access-gcbzb\") pod \"catalog-operator-68c6474976-5dnn6\" (UID: \"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.933692 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.944535 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxft9\" (UniqueName: \"kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9\") pod \"console-f9d7485db-6xxwt\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.960824 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.968326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ltzh\" (UniqueName: \"kubernetes.io/projected/2ab99367-5422-43bd-97d9-550e989d7542-kube-api-access-5ltzh\") pod \"dns-operator-744455d44c-6gl7z\" (UID: \"2ab99367-5422-43bd-97d9-550e989d7542\") " pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.968410 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz"] Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.969215 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.988800 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97nkc\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-kube-api-access-97nkc\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:48 crc kubenswrapper[4867]: I1212 06:50:48.994726 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.002817 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.010860 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtvsc\" (UniqueName: \"kubernetes.io/projected/45a95b26-1418-4107-bd27-0f67e3e6b9db-kube-api-access-rtvsc\") pod \"console-operator-58897d9998-pjj4n\" (UID: \"45a95b26-1418-4107-bd27-0f67e3e6b9db\") " pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.022833 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-592qx"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.026187 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7wk\" (UniqueName: \"kubernetes.io/projected/2a2028ce-f6e9-414f-b071-301f03fa6337-kube-api-access-tf7wk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d6wt4\" (UID: \"2a2028ce-f6e9-414f-b071-301f03fa6337\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.044676 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.052145 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9be1f196-2819-4290-a6df-3e0c3e9f0d3e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fh8gz\" (UID: \"9be1f196-2819-4290-a6df-3e0c3e9f0d3e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.057817 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7zlz5"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.067790 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbmt\" (UniqueName: \"kubernetes.io/projected/9aab65fa-d90b-489e-8443-07cac98cb8ee-kube-api-access-ggbmt\") pod \"apiserver-7bbb656c7d-72qzg\" (UID: \"9aab65fa-d90b-489e-8443-07cac98cb8ee\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.093780 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.105524 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.110050 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.116269 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jdj\" (UniqueName: \"kubernetes.io/projected/db9be553-1a67-4dc9-874b-7b79d332fcce-kube-api-access-79jdj\") pod \"service-ca-9c57cc56f-5r5vl\" (UID: \"db9be553-1a67-4dc9-874b-7b79d332fcce\") " pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.124179 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.124832 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qn79\" (UniqueName: \"kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79\") pod \"collect-profiles-29425365-5s62w\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.161881 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlth\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-kube-api-access-gdlth\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.162534 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9rzz\" (UniqueName: \"kubernetes.io/projected/f1ad2733-54b4-4365-9651-c6dbb04d7493-kube-api-access-m9rzz\") pod \"multus-admission-controller-857f4d67dd-lk4nq\" (UID: \"f1ad2733-54b4-4365-9651-c6dbb04d7493\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.182958 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjcdp\" (UniqueName: \"kubernetes.io/projected/93f987df-943c-45a9-aaf8-a0430358ba7f-kube-api-access-mjcdp\") pod \"machine-config-operator-74547568cd-s66s9\" (UID: \"93f987df-943c-45a9-aaf8-a0430358ba7f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.190177 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.211071 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28mcc\" (UniqueName: \"kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc\") pod \"controller-manager-879f6c89f-45697\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.224718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.225128 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c1e11f9-9309-4412-bda9-8d39d3192749-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ps2p4\" (UID: \"9c1e11f9-9309-4412-bda9-8d39d3192749\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.240144 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.243302 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9wrx\" (UniqueName: \"kubernetes.io/projected/865aeb86-6c89-4b2b-88c6-4d9637442caa-kube-api-access-r9wrx\") pod \"openshift-config-operator-7777fb866f-ztz7d\" (UID: \"865aeb86-6c89-4b2b-88c6-4d9637442caa\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.252127 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.261023 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.263382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjgjz\" (UniqueName: \"kubernetes.io/projected/818edab4-a66c-4739-bead-9d42292e468a-kube-api-access-cjgjz\") pod \"kube-storage-version-migrator-operator-b67b599dd-vkl2j\" (UID: \"818edab4-a66c-4739-bead-9d42292e468a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.278554 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.286162 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.287437 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn59v\" (UniqueName: \"kubernetes.io/projected/894bf2b1-e262-41f9-94d1-b2a2e8e8b86e-kube-api-access-rn59v\") pod \"openshift-apiserver-operator-796bbdcf4f-v4nks\" (UID: \"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.303066 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmnl\" (UniqueName: \"kubernetes.io/projected/4f7a80f9-c555-4325-b747-1a5a6e39e6ce-kube-api-access-gnmnl\") pod \"ingress-canary-nfwlh\" (UID: \"4f7a80f9-c555-4325-b747-1a5a6e39e6ce\") " pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.307975 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfk5k"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.308257 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.322039 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw5wb\" (UniqueName: \"kubernetes.io/projected/a8091880-bdd4-4a8d-8078-dec177dd01bd-kube-api-access-zw5wb\") pod \"machine-config-controller-84d6567774-w6q4w\" (UID: \"a8091880-bdd4-4a8d-8078-dec177dd01bd\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.360032 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.361980 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc4a337f-524f-40d0-a5d3-4f5653a2c5cd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sp4sz\" (UID: \"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.389042 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.391204 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.413413 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-nvs52"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.460181 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6gl7z"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.515614 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nmwfj"] Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.522812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lrj5h" event={"ID":"bd04d7ab-8933-4815-97ec-7be8b2e643ce","Type":"ContainerStarted","Data":"872d17f73f468cde472374124cc9623bad487c3351fe09b7498900baaf7fcc09"} Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.523866 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" event={"ID":"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a","Type":"ContainerStarted","Data":"d7c6d9c7fea8cb87e718f75405fdcabe36bb3551146aa749f458ab7c123eb470"} Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.525017 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" event={"ID":"5265271c-7232-4ce2-b0aa-fbfe5330e267","Type":"ContainerStarted","Data":"1ce3838ebac2635ffc3678b7eaa4d4e10768920d7f2c9e671e0b6fbc898d39eb"} Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.975305 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.975766 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.976979 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" Dec 12 06:50:49 crc kubenswrapper[4867]: E1212 06:50:49.977386 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.477369316 +0000 UTC m=+138.048750585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.979419 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.981577 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e33a71af-948c-4fe9-a374-d68bbb4fc862-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8qbr5\" (UID: \"e33a71af-948c-4fe9-a374-d68bbb4fc862\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.981870 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.984538 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.984814 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sd9l\" (UniqueName: \"kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l\") pod \"route-controller-manager-6576b87f9c-g9mhd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:49 crc kubenswrapper[4867]: I1212 06:50:49.992715 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nfwlh" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.018566 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078474 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078760 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078790 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-serving-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078814 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-audit-dir\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078837 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078926 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-node-pullsecrets\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.078948 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9gkq\" (UniqueName: \"kubernetes.io/projected/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-kube-api-access-v9gkq\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.079096 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wsf6\" (UniqueName: \"kubernetes.io/projected/81885549-e08a-4636-9f50-fb5116790c37-kube-api-access-9wsf6\") pod \"migrator-59844c95c7-njxxd\" (UID: \"81885549-e08a-4636-9f50-fb5116790c37\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.079136 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.079665 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.579639421 +0000 UTC m=+138.151020750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.079969 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-node-bootstrap-token\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080301 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smxv5\" (UniqueName: \"kubernetes.io/projected/3d58b26a-54e5-4699-9a2c-e4af556a4405-kube-api-access-smxv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080497 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080526 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080600 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxs6q\" (UniqueName: \"kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080637 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080660 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6mlb\" (UniqueName: \"kubernetes.io/projected/fa5560bb-1853-410e-8b95-42f498b3dbef-kube-api-access-b6mlb\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080830 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080866 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-certs\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080900 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.080959 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.081003 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-image-import-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.081094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svbfg\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.081424 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d58b26a-54e5-4699-9a2c-e4af556a4405-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.081458 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-srv-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.082563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-webhook-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083071 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-client\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083271 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-encryption-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083312 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-serving-cert\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc4s8\" (UniqueName: \"kubernetes.io/projected/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-kube-api-access-qc4s8\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083764 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.083796 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-audit\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.085515 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.085649 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4srp\" (UniqueName: \"kubernetes.io/projected/b0a26c08-980a-4ee4-b11a-a949232a50ae-kube-api-access-h4srp\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.086213 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.086472 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b0a26c08-980a-4ee4-b11a-a949232a50ae-tmpfs\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.086540 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmvd\" (UniqueName: \"kubernetes.io/projected/12a7c8be-2580-4a95-a465-eaac8f373729-kube-api-access-vsmvd\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.088619 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188146 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d58b26a-54e5-4699-9a2c-e4af556a4405-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188508 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-srv-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188527 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-webhook-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188545 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-client\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188595 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-encryption-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188619 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-metrics-tls\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188636 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-serving-cert\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcgqh\" (UniqueName: \"kubernetes.io/projected/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-kube-api-access-bcgqh\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188691 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc4s8\" (UniqueName: \"kubernetes.io/projected/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-kube-api-access-qc4s8\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188725 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188761 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-audit\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188822 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-mountpoint-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188859 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4srp\" (UniqueName: \"kubernetes.io/projected/b0a26c08-980a-4ee4-b11a-a949232a50ae-kube-api-access-h4srp\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188876 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-socket-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188889 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-csi-data-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188929 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188950 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b0a26c08-980a-4ee4-b11a-a949232a50ae-tmpfs\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188968 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmvd\" (UniqueName: \"kubernetes.io/projected/12a7c8be-2580-4a95-a465-eaac8f373729-kube-api-access-vsmvd\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-plugins-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.188999 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knsdb\" (UniqueName: \"kubernetes.io/projected/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-kube-api-access-knsdb\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189069 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189087 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-serving-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189389 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-audit-dir\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189410 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-audit-dir\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189436 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189459 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189492 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-node-pullsecrets\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.189524 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9gkq\" (UniqueName: \"kubernetes.io/projected/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-kube-api-access-v9gkq\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.190596 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.190708 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.191132 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-serving-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.191633 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b0a26c08-980a-4ee4-b11a-a949232a50ae-tmpfs\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.192046 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-audit\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193269 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-encryption-config\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193552 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-serving-cert\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193592 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/12a7c8be-2580-4a95-a465-eaac8f373729-node-pullsecrets\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193727 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-apiservice-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193811 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.193884 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-config-volume\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194117 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194402 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wsf6\" (UniqueName: \"kubernetes.io/projected/81885549-e08a-4636-9f50-fb5116790c37-kube-api-access-9wsf6\") pod \"migrator-59844c95c7-njxxd\" (UID: \"81885549-e08a-4636-9f50-fb5116790c37\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194531 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194580 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-node-bootstrap-token\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194821 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smxv5\" (UniqueName: \"kubernetes.io/projected/3d58b26a-54e5-4699-9a2c-e4af556a4405-kube-api-access-smxv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194869 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-registration-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.194989 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195010 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195072 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxs6q\" (UniqueName: \"kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.195418 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.695406379 +0000 UTC m=+138.266787648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195443 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195466 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6mlb\" (UniqueName: \"kubernetes.io/projected/fa5560bb-1853-410e-8b95-42f498b3dbef-kube-api-access-b6mlb\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195762 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195784 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-certs\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195860 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195924 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svbfg\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.195948 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-image-import-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.196155 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.197759 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0a26c08-980a-4ee4-b11a-a949232a50ae-webhook-cert\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.198245 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-profile-collector-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.198263 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.198446 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.198457 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.199172 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-node-bootstrap-token\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.199174 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12a7c8be-2580-4a95-a465-eaac8f373729-etcd-client\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.199301 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/12a7c8be-2580-4a95-a465-eaac8f373729-image-import-ca\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.199350 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.199961 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fa5560bb-1853-410e-8b95-42f498b3dbef-certs\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.218095 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.228559 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4srp\" (UniqueName: \"kubernetes.io/projected/b0a26c08-980a-4ee4-b11a-a949232a50ae-kube-api-access-h4srp\") pod \"packageserver-d55dfcdfc-77bzq\" (UID: \"b0a26c08-980a-4ee4-b11a-a949232a50ae\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.228949 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-srv-cert\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.229139 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmvd\" (UniqueName: \"kubernetes.io/projected/12a7c8be-2580-4a95-a465-eaac8f373729-kube-api-access-vsmvd\") pod \"apiserver-76f77b778f-c7vxj\" (UID: \"12a7c8be-2580-4a95-a465-eaac8f373729\") " pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.232243 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3d58b26a-54e5-4699-9a2c-e4af556a4405-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.247000 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc4s8\" (UniqueName: \"kubernetes.io/projected/2bc83e9c-4e30-45b0-aefe-bd648a8aefe4-kube-api-access-qc4s8\") pod \"package-server-manager-789f6589d5-ngkbp\" (UID: \"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.249110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wsf6\" (UniqueName: \"kubernetes.io/projected/81885549-e08a-4636-9f50-fb5116790c37-kube-api-access-9wsf6\") pod \"migrator-59844c95c7-njxxd\" (UID: \"81885549-e08a-4636-9f50-fb5116790c37\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.254805 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.255450 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9gkq\" (UniqueName: \"kubernetes.io/projected/a11290f4-5f18-4a1d-8f5b-4cb5b27351fa-kube-api-access-v9gkq\") pod \"olm-operator-6b444d44fb-d6p85\" (UID: \"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.267589 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.286505 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxs6q\" (UniqueName: \"kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q\") pod \"marketplace-operator-79b997595-qlb8v\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.290081 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6mlb\" (UniqueName: \"kubernetes.io/projected/fa5560bb-1853-410e-8b95-42f498b3dbef-kube-api-access-b6mlb\") pod \"machine-config-server-ht7nj\" (UID: \"fa5560bb-1853-410e-8b95-42f498b3dbef\") " pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.299859 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.300148 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.800116723 +0000 UTC m=+138.371497992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300187 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-config-volume\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300232 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300762 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-registration-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300816 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-metrics-tls\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300837 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcgqh\" (UniqueName: \"kubernetes.io/projected/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-kube-api-access-bcgqh\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.300872 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.80085571 +0000 UTC m=+138.372236979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300929 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-mountpoint-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300963 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-socket-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300999 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-csi-data-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-plugins-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301068 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knsdb\" (UniqueName: \"kubernetes.io/projected/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-kube-api-access-knsdb\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301113 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-registration-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301313 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-socket-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301391 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-csi-data-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301448 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-plugins-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.300887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-config-volume\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.301754 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-mountpoint-dir\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.319695 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smxv5\" (UniqueName: \"kubernetes.io/projected/3d58b26a-54e5-4699-9a2c-e4af556a4405-kube-api-access-smxv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-n64dp\" (UID: \"3d58b26a-54e5-4699-9a2c-e4af556a4405\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.322092 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-metrics-tls\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.326328 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.333437 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.336197 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svbfg\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.340767 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.346652 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lk4nq"] Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.359834 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.362441 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.369781 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcgqh\" (UniqueName: \"kubernetes.io/projected/b3f15d74-7f84-4f87-8b11-7ec1f993f3ef-kube-api-access-bcgqh\") pod \"csi-hostpathplugin-t8lml\" (UID: \"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef\") " pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.384244 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.388210 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knsdb\" (UniqueName: \"kubernetes.io/projected/22c4c88c-1195-4f05-b3b4-f1308c7d5e68-kube-api-access-knsdb\") pod \"dns-default-fzxx5\" (UID: \"22c4c88c-1195-4f05-b3b4-f1308c7d5e68\") " pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.390456 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.400621 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ht7nj" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.402948 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.403009 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.902992322 +0000 UTC m=+138.474373601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.403962 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.404398 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:50.904383685 +0000 UTC m=+138.475764954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.411690 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.429202 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" Dec 12 06:50:50 crc kubenswrapper[4867]: W1212 06:50:50.435810 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1ad2733_54b4_4365_9651_c6dbb04d7493.slice/crio-dcf78310e7285c3d4f1937d2aa4cac512e5cfb144611a300645364a66faf3981 WatchSource:0}: Error finding container dcf78310e7285c3d4f1937d2aa4cac512e5cfb144611a300645364a66faf3981: Status 404 returned error can't find the container with id dcf78310e7285c3d4f1937d2aa4cac512e5cfb144611a300645364a66faf3981 Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.504700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.505024 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.005001651 +0000 UTC m=+138.576382920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.575244 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lrj5h" event={"ID":"bd04d7ab-8933-4815-97ec-7be8b2e643ce","Type":"ContainerStarted","Data":"676bbddd6826acb8be96a64fc9068a3a853900bdee798124e54f3af90e7f52d7"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.578815 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" event={"ID":"8c966cee-93bd-41b1-885c-e46c0f9ede1a","Type":"ContainerStarted","Data":"a8b9376b0024167246e1ef55c941c610ae1cb31976ddcedd703850112c9c53a6"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.581319 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" event={"ID":"5265271c-7232-4ce2-b0aa-fbfe5330e267","Type":"ContainerStarted","Data":"62bcbf046a6ad4a161b6679543f4e1a1f7927d3d29f4a5c38f0cdda8cfd5c1b1"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.584782 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" event={"ID":"eecb4b3a-68a9-4e82-8d06-2f06fff6d104","Type":"ContainerStarted","Data":"2ae7a306424ce850453d76c26c47740005a4ce029e4013baa5f0e6fa213213b4"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.588189 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" event={"ID":"2ab99367-5422-43bd-97d9-550e989d7542","Type":"ContainerStarted","Data":"4a85cbf15cf1686aaf24cef0b173eb752ecc90afe09426cb36fae8dba9835761"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.591185 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nmwfj" event={"ID":"a648d31a-4689-434c-b89b-c0d96065ce10","Type":"ContainerStarted","Data":"23cc77d5863d45157b434d4143594d46b1e6045de3680bed79f895a68a367297"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.593540 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" event={"ID":"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a","Type":"ContainerStarted","Data":"8e2a0d9682e80b0e7b8c5bc0fc85162ba270ce0f0a3bf54792e09bb108041fee"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.606040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.607317 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.107301257 +0000 UTC m=+138.678682526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.613528 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" event={"ID":"e3ae7c8f-1ae7-443f-a049-7eca845b9387","Type":"ContainerStarted","Data":"8d9fe4cd3f9fd0eb876baa795b712ccf0720ba93187627d8aeff2eee832b7721"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.618076 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" event={"ID":"b9ce7d0e-4827-492c-a71e-53299557a2b5","Type":"ContainerStarted","Data":"f8b3d1c529a6a53f4a4681634073b3cb6dd5e78e3c0b6bb4b631aae6b730ec29"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.710674 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.711157 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.211139029 +0000 UTC m=+138.782520298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.735532 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" event={"ID":"44957455-bf87-4493-bee7-0304b2165061","Type":"ContainerStarted","Data":"230c02a30bd0fad0e98dcfa2a0166e2e32d58f298d527b11eccabb157fb7cdf8"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.746162 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" event={"ID":"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61","Type":"ContainerStarted","Data":"bf59bed22ba69a6d9cb4afbedc5bdfe62c5cbaa879d466682c43f5c74231fcc0"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.746234 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" event={"ID":"bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61","Type":"ContainerStarted","Data":"a5bb4b29fff74a686938899d8d64a2efa058cbbdff49679fe4e080dad91ab3e6"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.746373 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.754944 4867 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5dnn6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.754983 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" podUID="bc72dbc7-9ee0-4d26-8dcf-88e9b8b9da61" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.758994 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" event={"ID":"f1ad2733-54b4-4365-9651-c6dbb04d7493","Type":"ContainerStarted","Data":"dcf78310e7285c3d4f1937d2aa4cac512e5cfb144611a300645364a66faf3981"} Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.767945 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg"] Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.812484 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.812876 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.312855721 +0000 UTC m=+138.884237000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.913858 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:50 crc kubenswrapper[4867]: E1212 06:50:50.914689 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.414670605 +0000 UTC m=+138.986051874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.935739 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:50 crc kubenswrapper[4867]: W1212 06:50:50.976642 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aab65fa_d90b_489e_8443_07cac98cb8ee.slice/crio-34e0e0bc514fc97cf30108d206e39a027722b55d38828fda2a898175896f65e2 WatchSource:0}: Error finding container 34e0e0bc514fc97cf30108d206e39a027722b55d38828fda2a898175896f65e2: Status 404 returned error can't find the container with id 34e0e0bc514fc97cf30108d206e39a027722b55d38828fda2a898175896f65e2 Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.980438 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:50 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:50 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:50 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:50 crc kubenswrapper[4867]: I1212 06:50:50.980491 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.016529 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.016850 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.516838788 +0000 UTC m=+139.088220057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.118103 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.118594 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.61856366 +0000 UTC m=+139.189944929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.125380 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.126084 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.625871534 +0000 UTC m=+139.197252803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.186546 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.203419 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.222952 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.228203 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjj4n"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.228794 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.229450 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.72943042 +0000 UTC m=+139.300811689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.229618 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.251373 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lrj5h" podStartSLOduration=121.251355435 podStartE2EDuration="2m1.251355435s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.249575613 +0000 UTC m=+138.820956872" watchObservedRunningTime="2025-12-12 06:50:51.251355435 +0000 UTC m=+138.822736704" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.290256 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-592qx" podStartSLOduration=121.290241225 podStartE2EDuration="2m1.290241225s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.289284462 +0000 UTC m=+138.860665731" watchObservedRunningTime="2025-12-12 06:50:51.290241225 +0000 UTC m=+138.861622494" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.330496 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" podStartSLOduration=120.328218043 podStartE2EDuration="2m0.328218043s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.325274642 +0000 UTC m=+138.896655921" watchObservedRunningTime="2025-12-12 06:50:51.328218043 +0000 UTC m=+138.899599312" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.332148 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.332551 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.832537816 +0000 UTC m=+139.403919085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.360593 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" podStartSLOduration=120.360575616 podStartE2EDuration="2m0.360575616s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.359520211 +0000 UTC m=+138.930901480" watchObservedRunningTime="2025-12-12 06:50:51.360575616 +0000 UTC m=+138.931956885" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.433588 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.433976 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:51.93395524 +0000 UTC m=+139.505336519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.534688 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.535204 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.035192211 +0000 UTC m=+139.606573480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.584694 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.605331 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.611372 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz"] Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.612519 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:50:51 crc kubenswrapper[4867]: W1212 06:50:51.632377 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod894bf2b1_e262_41f9_94d1_b2a2e8e8b86e.slice/crio-4e1c3a275b959cdab59dc9f00cd0a9e487d8f0a910216027f448cff63dec1455 WatchSource:0}: Error finding container 4e1c3a275b959cdab59dc9f00cd0a9e487d8f0a910216027f448cff63dec1455: Status 404 returned error can't find the container with id 4e1c3a275b959cdab59dc9f00cd0a9e487d8f0a910216027f448cff63dec1455 Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.635649 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.635937 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.13592001 +0000 UTC m=+139.707301279 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.737341 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.737705 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.237692142 +0000 UTC m=+139.809073411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.804463 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" event={"ID":"8c966cee-93bd-41b1-885c-e46c0f9ede1a","Type":"ContainerStarted","Data":"ec95f6abf8f5cbab7c5bf3b6f77076d5cb80a96c6329f094bdaff3c0be175b4f"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.805209 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.841988 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.842303 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.342287633 +0000 UTC m=+139.913668902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.842427 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" event={"ID":"2ab99367-5422-43bd-97d9-550e989d7542","Type":"ContainerStarted","Data":"2d0cc0ca16e5e535e0f7e3f2da845738b3f3d68e295d411863f30515c2aaa5b0"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.844034 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" podStartSLOduration=121.844011804 podStartE2EDuration="2m1.844011804s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.841532255 +0000 UTC m=+139.412913524" watchObservedRunningTime="2025-12-12 06:50:51.844011804 +0000 UTC m=+139.415393073" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.849152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" event={"ID":"3dc551d1-7e53-4bb5-80a7-4a9f30ce5e0a","Type":"ContainerStarted","Data":"4b906f505a2552af9d6bfda983ccb13203b45b6a1f2189bb22f1b6eb9765f8ca"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.873848 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6xxwt" event={"ID":"759c6366-d32d-47d9-89c4-ce7d274637c5","Type":"ContainerStarted","Data":"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.873893 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6xxwt" event={"ID":"759c6366-d32d-47d9-89c4-ce7d274637c5","Type":"ContainerStarted","Data":"80245644461f06327bd42133e8da9b5841bf7dac272aa948a8abb04a5526d616"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.885476 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" event={"ID":"9be1f196-2819-4290-a6df-3e0c3e9f0d3e","Type":"ContainerStarted","Data":"c2e1df987990c1ff101c199b5ad7cf0acb5f5f40290727b3dab5b510783faf1a"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.891766 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" event={"ID":"e3ae7c8f-1ae7-443f-a049-7eca845b9387","Type":"ContainerStarted","Data":"1ec79ff14d31a8eaacd802c629e864db3fe4aacbc802a5d4711211004fd3ccf1"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.897956 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" event={"ID":"9aab65fa-d90b-489e-8443-07cac98cb8ee","Type":"ContainerStarted","Data":"34e0e0bc514fc97cf30108d206e39a027722b55d38828fda2a898175896f65e2"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.906814 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ppf4t" podStartSLOduration=121.906796405 podStartE2EDuration="2m1.906796405s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.865105838 +0000 UTC m=+139.436487107" watchObservedRunningTime="2025-12-12 06:50:51.906796405 +0000 UTC m=+139.478177674" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.910894 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-nvs52" event={"ID":"44957455-bf87-4493-bee7-0304b2165061","Type":"ContainerStarted","Data":"4b3e6962e368032e5606b16757ba5a7ad0ea04701df01e56a57b4e467d5b4ad9"} Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.935792 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fjccs" podStartSLOduration=121.935775639 podStartE2EDuration="2m1.935775639s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.93290873 +0000 UTC m=+139.504290019" watchObservedRunningTime="2025-12-12 06:50:51.935775639 +0000 UTC m=+139.507156908" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.948396 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6xxwt" podStartSLOduration=121.94837425 podStartE2EDuration="2m1.94837425s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:51.908699331 +0000 UTC m=+139.480080610" watchObservedRunningTime="2025-12-12 06:50:51.94837425 +0000 UTC m=+139.519755509" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.948967 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:51 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:51 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:51 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.949019 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.950946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:51 crc kubenswrapper[4867]: E1212 06:50:51.956790 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.456772471 +0000 UTC m=+140.028153800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:51 crc kubenswrapper[4867]: I1212 06:50:51.979526 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5r5vl"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.000281 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.014014 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.016321 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.019063 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" event={"ID":"eecb4b3a-68a9-4e82-8d06-2f06fff6d104","Type":"ContainerStarted","Data":"8bfcfd4e342e31bb99b145ef00abb9793a1e89013377b5e372d22ca5f863c0bf"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.019524 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" event={"ID":"eecb4b3a-68a9-4e82-8d06-2f06fff6d104","Type":"ContainerStarted","Data":"4ed67a330a54129a2d2a71c97f1a20f78abbe81de3d50c09f91ddbffbfc7b7aa"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.021067 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c7vxj"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.024177 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nfwlh"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.053686 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.055856 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.555830488 +0000 UTC m=+140.127211757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.058524 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.062595 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.083447 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" event={"ID":"f1ad2733-54b4-4365-9651-c6dbb04d7493","Type":"ContainerStarted","Data":"2462eb0bcfea0f0f80fd9c3c5c3e96f8b8fc12a4a03181608e16918dfb5b3f57"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.091969 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-lqkgz" podStartSLOduration=122.091950552 podStartE2EDuration="2m2.091950552s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.090325053 +0000 UTC m=+139.661706322" watchObservedRunningTime="2025-12-12 06:50:52.091950552 +0000 UTC m=+139.663331821" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.099778 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" event={"ID":"f8f8fec8-9201-4083-8948-6c2958ef5ec7","Type":"ContainerStarted","Data":"9e2110d497e4563b3c8332e160c4fc5b1518fa3a31c4dcbcbbc4b73a291dec06"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.115792 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.127800 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.129830 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-52qk7"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.155088 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.155462 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.655450291 +0000 UTC m=+140.226831560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.199064 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nmwfj" event={"ID":"a648d31a-4689-434c-b89b-c0d96065ce10","Type":"ContainerStarted","Data":"af5f860d102e4fdc1ee85624e77ee497c08d1933f029ae7dd2d2f6759a619396"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.200093 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.212440 4867 patch_prober.go:28] interesting pod/downloads-7954f5f757-nmwfj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.212497 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nmwfj" podUID="a648d31a-4689-434c-b89b-c0d96065ce10" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.220339 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" event={"ID":"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd","Type":"ContainerStarted","Data":"16c59cefb5421ea013368e70b0d2faebb027750561353505632a0ab45da0a8dc"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.225818 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-nmwfj" podStartSLOduration=122.225801423 podStartE2EDuration="2m2.225801423s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.224925071 +0000 UTC m=+139.796306340" watchObservedRunningTime="2025-12-12 06:50:52.225801423 +0000 UTC m=+139.797182692" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.239269 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.241331 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fzxx5"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.251888 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" event={"ID":"865aeb86-6c89-4b2b-88c6-4d9637442caa","Type":"ContainerStarted","Data":"d5c90bf20881426041a0518f65a7d33805a5af28e5c44ea7225f7df5c9732b46"} Dec 12 06:50:52 crc kubenswrapper[4867]: W1212 06:50:52.252540 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32c762a6_9bfc_40d6_8a78_eb4189e885bd.slice/crio-a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe WatchSource:0}: Error finding container a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe: Status 404 returned error can't find the container with id a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.256047 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.256720 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.756704361 +0000 UTC m=+140.328085620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.256846 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.257137 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.757102431 +0000 UTC m=+140.328483700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.270926 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t8lml"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.270959 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.311092 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.319412 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" event={"ID":"45a95b26-1418-4107-bd27-0f67e3e6b9db","Type":"ContainerStarted","Data":"0f00ea3b388d1205b82e67f15ae90e5deb8a7596a3cdec0a62f8874d399720d8"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.320098 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.321284 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.329561 4867 patch_prober.go:28] interesting pod/console-operator-58897d9998-pjj4n container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.329600 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" podUID="45a95b26-1418-4107-bd27-0f67e3e6b9db" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.331785 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.335371 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" event={"ID":"b9ce7d0e-4827-492c-a71e-53299557a2b5","Type":"ContainerStarted","Data":"26cec27460607c4555a195ff8b19087d658140a6b82e66db17e46936f72a86ab"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.335409 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" event={"ID":"b9ce7d0e-4827-492c-a71e-53299557a2b5","Type":"ContainerStarted","Data":"4c50391c6d34d7504b5ef661acf958d334f78b5c0fe6d207e04ccdf1c432f34c"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.379318 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.379608 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.879585809 +0000 UTC m=+140.450967079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.380355 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" event={"ID":"37ad992d-0e36-434e-9d11-106ef955c5dd","Type":"ContainerStarted","Data":"2f27a33585dc01956d4da1e55a919e4a27bca56559f9a5aad5f8fd52418450f9"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.381009 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.382998 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.383577 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.883565175 +0000 UTC m=+140.454946444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.384589 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" podStartSLOduration=122.384557558 podStartE2EDuration="2m2.384557558s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.34282691 +0000 UTC m=+139.914208179" watchObservedRunningTime="2025-12-12 06:50:52.384557558 +0000 UTC m=+139.955938827" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.384716 4867 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-g9mhd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.386622 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.400292 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7zlz5" podStartSLOduration=121.400270203 podStartE2EDuration="2m1.400270203s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.391086454 +0000 UTC m=+139.962467713" watchObservedRunningTime="2025-12-12 06:50:52.400270203 +0000 UTC m=+139.971651472" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.401056 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.409062 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" event={"ID":"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e","Type":"ContainerStarted","Data":"4e1c3a275b959cdab59dc9f00cd0a9e487d8f0a910216027f448cff63dec1455"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.450195 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" event={"ID":"9c1e11f9-9309-4412-bda9-8d39d3192749","Type":"ContainerStarted","Data":"84f00ab1751760bd48eeb1a18e7273cda03cebd49f0c98aee42f559c8cdd0979"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.463422 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ht7nj" event={"ID":"fa5560bb-1853-410e-8b95-42f498b3dbef","Type":"ContainerStarted","Data":"1b3f64a4d0e0dffc9bea286a039e9b89eea8efdf0e3198c455d6d47cd2821718"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.463460 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ht7nj" event={"ID":"fa5560bb-1853-410e-8b95-42f498b3dbef","Type":"ContainerStarted","Data":"f7ab6bd37da2dd70e39c359b060d006a3a67c41b2e72466d4a0dd3a72c73257d"} Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.477072 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" podStartSLOduration=122.477033789 podStartE2EDuration="2m2.477033789s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.461550649 +0000 UTC m=+140.032931918" watchObservedRunningTime="2025-12-12 06:50:52.477033789 +0000 UTC m=+140.048415058" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.477530 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" podStartSLOduration=121.477525761 podStartE2EDuration="2m1.477525761s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.430719072 +0000 UTC m=+140.002100341" watchObservedRunningTime="2025-12-12 06:50:52.477525761 +0000 UTC m=+140.048907030" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.485253 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5dnn6" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.485764 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.486066 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.986046555 +0000 UTC m=+140.557427824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.486184 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.488997 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:52.988926514 +0000 UTC m=+140.560307783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.532623 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" podStartSLOduration=122.532605338 podStartE2EDuration="2m2.532605338s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.530713573 +0000 UTC m=+140.102094852" watchObservedRunningTime="2025-12-12 06:50:52.532605338 +0000 UTC m=+140.103986607" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.545922 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.546850 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.563046 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.576424 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.577870 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ht7nj" podStartSLOduration=6.577837439 podStartE2EDuration="6.577837439s" podCreationTimestamp="2025-12-12 06:50:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:52.572316797 +0000 UTC m=+140.143698066" watchObservedRunningTime="2025-12-12 06:50:52.577837439 +0000 UTC m=+140.149218708" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.587013 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.587197 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.087158482 +0000 UTC m=+140.658539751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.587487 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.589953 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.089944409 +0000 UTC m=+140.661325678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.689044 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.689236 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.189195182 +0000 UTC m=+140.760576451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.689535 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zg4m\" (UniqueName: \"kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.689591 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.689638 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.689666 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.689962 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.189950679 +0000 UTC m=+140.761331938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.734722 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.746691 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.746832 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.751669 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.791347 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.791572 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zg4m\" (UniqueName: \"kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.791677 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.791710 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.792119 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.792268 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.792748 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.292722387 +0000 UTC m=+140.864103676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.821271 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zg4m\" (UniqueName: \"kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m\") pod \"certified-operators-g678h\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.899885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.900311 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.900363 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.900400 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54nv8\" (UniqueName: \"kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:52 crc kubenswrapper[4867]: E1212 06:50:52.900693 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.400681898 +0000 UTC m=+140.972063167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.940437 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.941674 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:52 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:52 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:52 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.941714 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.957776 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:52 crc kubenswrapper[4867]: I1212 06:50:52.964307 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.022766 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.024980 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.025114 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.525098433 +0000 UTC m=+141.096479702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.025282 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.025306 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.025345 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.025382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54nv8\" (UniqueName: \"kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.025843 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.52583594 +0000 UTC m=+141.097217209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.026459 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.026588 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.059056 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54nv8\" (UniqueName: \"kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8\") pod \"community-operators-fmxsr\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.126284 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.126536 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.126777 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.126894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrztq\" (UniqueName: \"kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.127024 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.62700799 +0000 UTC m=+141.198389259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.139092 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.142718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.163798 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238079 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238322 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238361 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238380 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238404 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238434 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gml9m\" (UniqueName: \"kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.238454 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrztq\" (UniqueName: \"kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.239345 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.239576 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.73956631 +0000 UTC m=+141.310947579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.239962 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.279865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrztq\" (UniqueName: \"kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq\") pod \"certified-operators-ncppm\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.311639 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.339118 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.339414 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.339448 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.339505 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.839482849 +0000 UTC m=+141.410864108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.339573 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gml9m\" (UniqueName: \"kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.339799 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.351170 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.399261 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gml9m\" (UniqueName: \"kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m\") pod \"community-operators-rkrfw\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.417504 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.440826 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.441133 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:53.941121539 +0000 UTC m=+141.512502808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.469478 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.521054 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" event={"ID":"93f987df-943c-45a9-aaf8-a0430358ba7f","Type":"ContainerStarted","Data":"a56504b8118fcfe5ed541237afc78f297823292bb62c56f56011358e9856b1a1"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.523760 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.541646 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.542101 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.042079983 +0000 UTC m=+141.613461262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.567727 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerStarted","Data":"d2ec66ee7b435fd6bc7e6a6cece258be1b306796892c12aa31d1cc0b1cfb1c33"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.622835 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" event={"ID":"b0a26c08-980a-4ee4-b11a-a949232a50ae","Type":"ContainerStarted","Data":"66e5b24f9cfc05ee0bb70cc4257c9fe3d9648d59cc14153866875f888f48a23d"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.623261 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.623291 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" event={"ID":"b0a26c08-980a-4ee4-b11a-a949232a50ae","Type":"ContainerStarted","Data":"ca25c76049c035ce4fa8da7d86f38d9970b7ea2158cc12926b83c8131ece3e1c"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.624861 4867 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-77bzq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.624909 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" podUID="b0a26c08-980a-4ee4-b11a-a949232a50ae" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.643361 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.643722 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.143709243 +0000 UTC m=+141.715090512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.653868 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" event={"ID":"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4","Type":"ContainerStarted","Data":"e30fac2aeb2f8b809733e0419904a73a4b4476ea03795fe3affcfb3f3cf0782a"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.653929 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" event={"ID":"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4","Type":"ContainerStarted","Data":"ec251865a157c937ee454da8d2984bbb808048c8725caff7ecd7c7b4f5e4db66"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.683945 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" event={"ID":"2a2028ce-f6e9-414f-b071-301f03fa6337","Type":"ContainerStarted","Data":"7fe32ebfcf945a2631819987fc72cdbe45e4db2d528200a352905a169f3cb58a"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.683995 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" event={"ID":"2a2028ce-f6e9-414f-b071-301f03fa6337","Type":"ContainerStarted","Data":"0e454f124d2319b6a3753cd9caa40876fbde9fe8b5f43400839c7d43a6aaa451"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.697217 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" event={"ID":"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef","Type":"ContainerStarted","Data":"778a46e52accc587a6cbca670d4c4a110cf16ad892b800c91036389967d6362f"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.717565 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" event={"ID":"7b0920c8-ab26-4b72-b22b-c4123498a461","Type":"ContainerStarted","Data":"80ae9ff540d43636b6b2f009be7268e34af00509ce725d3b73f7a715580f793c"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.745403 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.746762 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.246739496 +0000 UTC m=+141.818120765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.758212 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" event={"ID":"81885549-e08a-4636-9f50-fb5116790c37","Type":"ContainerStarted","Data":"ef2d4ee061f16723cbcc046b733a5424934ded60ef6bae737a0612a0548cc767"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.770377 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d6wt4" podStartSLOduration=123.7703493 podStartE2EDuration="2m3.7703493s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:53.757890562 +0000 UTC m=+141.329271831" watchObservedRunningTime="2025-12-12 06:50:53.7703493 +0000 UTC m=+141.341730569" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.770565 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" podStartSLOduration=122.770560005 podStartE2EDuration="2m2.770560005s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:53.668855334 +0000 UTC m=+141.240236603" watchObservedRunningTime="2025-12-12 06:50:53.770560005 +0000 UTC m=+141.341941274" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.814764 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12a7c8be_2580_4a95_a465_eaac8f373729.slice/crio-c2d56966322b911c87d923113f428049314a71dea926b58cc8bd31a9774e7a8b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12a7c8be_2580_4a95_a465_eaac8f373729.slice/crio-conmon-c2d56966322b911c87d923113f428049314a71dea926b58cc8bd31a9774e7a8b.scope\": RecentStats: unable to find data in memory cache]" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.852414 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.852822 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.352810792 +0000 UTC m=+141.924192061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.854587 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" event={"ID":"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa","Type":"ContainerStarted","Data":"41f893d4ddfe586bc4d09315f61adca636b4a551ba21588754e52b4c89727e62"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.854641 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" event={"ID":"a11290f4-5f18-4a1d-8f5b-4cb5b27351fa","Type":"ContainerStarted","Data":"d74e4a940894d897b584465a7044c65ee5ea0be7723229da83faacb746012920"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.855338 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.873454 4867 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-d6p85 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.873500 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" podUID="a11290f4-5f18-4a1d-8f5b-4cb5b27351fa" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.893495 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" podStartSLOduration=123.893479784 podStartE2EDuration="2m3.893479784s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:53.892108222 +0000 UTC m=+141.463489491" watchObservedRunningTime="2025-12-12 06:50:53.893479784 +0000 UTC m=+141.464861053" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.922563 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" event={"ID":"f1ad2733-54b4-4365-9651-c6dbb04d7493","Type":"ContainerStarted","Data":"d518644569a9eff6e5db048fc205f7a8a3bdefc5852b8e5bece682dfba0a07a7"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.944053 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" podStartSLOduration=122.944034793 podStartE2EDuration="2m2.944034793s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:53.941993835 +0000 UTC m=+141.513375104" watchObservedRunningTime="2025-12-12 06:50:53.944034793 +0000 UTC m=+141.515416062" Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.946315 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" event={"ID":"e33a71af-948c-4fe9-a374-d68bbb4fc862","Type":"ContainerStarted","Data":"9da44d22058a02cc4d04e2669ff3e72caf51764886b0b956e0920e6826c253e4"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.946360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" event={"ID":"e33a71af-948c-4fe9-a374-d68bbb4fc862","Type":"ContainerStarted","Data":"a3825b49dec00ecaecc4c858536212ba31a4f6243f7a8b414bb40185729c23c9"} Dec 12 06:50:53 crc kubenswrapper[4867]: I1212 06:50:53.960136 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:53 crc kubenswrapper[4867]: E1212 06:50:53.961929 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.461910051 +0000 UTC m=+142.033291320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.002762 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" event={"ID":"9be1f196-2819-4290-a6df-3e0c3e9f0d3e","Type":"ContainerStarted","Data":"1f314e3a58dd3b60b874fb7f4d967c165c2926ee8e04dcdece61075974faf40d"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.004201 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:54 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:54 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:54 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.004267 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.006106 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lk4nq" podStartSLOduration=123.006097156 podStartE2EDuration="2m3.006097156s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.004061228 +0000 UTC m=+141.575442497" watchObservedRunningTime="2025-12-12 06:50:54.006097156 +0000 UTC m=+141.577478425" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.038358 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8qbr5" podStartSLOduration=124.038343407 podStartE2EDuration="2m4.038343407s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.038107452 +0000 UTC m=+141.609488721" watchObservedRunningTime="2025-12-12 06:50:54.038343407 +0000 UTC m=+141.609724676" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.041473 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fzxx5" event={"ID":"22c4c88c-1195-4f05-b3b4-f1308c7d5e68","Type":"ContainerStarted","Data":"ccb2eda0b930d7e3ac13536a0a36ea3136e7193d60d574960c30d1ebb71a3471"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.062472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.064909 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.564890902 +0000 UTC m=+142.136272171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.092700 4867 generic.go:334] "Generic (PLEG): container finished" podID="865aeb86-6c89-4b2b-88c6-4d9637442caa" containerID="e0288aad80e219d9f6c6a4d2f21f9a517bc1a0538b169ef35a4c34b963066e5b" exitCode=0 Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.092804 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" event={"ID":"865aeb86-6c89-4b2b-88c6-4d9637442caa","Type":"ContainerStarted","Data":"4c8a4336de3e469b5039c5cf5e4bdc228d56fdb11c04b7a97dadbeab6423b546"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.116195 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" event={"ID":"865aeb86-6c89-4b2b-88c6-4d9637442caa","Type":"ContainerDied","Data":"e0288aad80e219d9f6c6a4d2f21f9a517bc1a0538b169ef35a4c34b963066e5b"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.116270 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.150872 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" event={"ID":"9c1e11f9-9309-4412-bda9-8d39d3192749","Type":"ContainerStarted","Data":"98de36a9cbb85210cf75184dd1e28a06c9ff07a410fb2c3edddc90e2d167f02b"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.150915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ps2p4" event={"ID":"9c1e11f9-9309-4412-bda9-8d39d3192749","Type":"ContainerStarted","Data":"6109dddd5961d3710e0cfd0cc8542746184ab4e1c25c6cf9b13db2c18eb73d37"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.160823 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" podStartSLOduration=124.160806436 podStartE2EDuration="2m4.160806436s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.159686789 +0000 UTC m=+141.731068058" watchObservedRunningTime="2025-12-12 06:50:54.160806436 +0000 UTC m=+141.732187705" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.161319 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fh8gz" podStartSLOduration=124.161314978 podStartE2EDuration="2m4.161314978s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.086735785 +0000 UTC m=+141.658117054" watchObservedRunningTime="2025-12-12 06:50:54.161314978 +0000 UTC m=+141.732696247" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.164802 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.165126 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.665100948 +0000 UTC m=+142.236482217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.166445 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" event={"ID":"db9be553-1a67-4dc9-874b-7b79d332fcce","Type":"ContainerStarted","Data":"6dfb35e59e35f5270fae9ea41a0d8921781024cde0bda702a0f815e46f6a572a"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.166479 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" event={"ID":"db9be553-1a67-4dc9-874b-7b79d332fcce","Type":"ContainerStarted","Data":"28a2642d5f40335e4a64ef120d2f30b426ca1f3ad5a3d41c165bc00dc38d0b7b"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.184683 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" event={"ID":"a8091880-bdd4-4a8d-8078-dec177dd01bd","Type":"ContainerStarted","Data":"8b06f37ea5bad2255e2a0b55d217952b032c2a32b2db9057f374e0a1c66f42eb"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.184725 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" event={"ID":"a8091880-bdd4-4a8d-8078-dec177dd01bd","Type":"ContainerStarted","Data":"ea6195c73ea6447b72078f83a4f13d20c10ffca8beef814101bb43f45514046f"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.196707 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5r5vl" podStartSLOduration=123.196690934 podStartE2EDuration="2m3.196690934s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.194671045 +0000 UTC m=+141.766052314" watchObservedRunningTime="2025-12-12 06:50:54.196690934 +0000 UTC m=+141.768072203" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.204381 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nfwlh" event={"ID":"4f7a80f9-c555-4325-b747-1a5a6e39e6ce","Type":"ContainerStarted","Data":"fc6cf76adac0d4a01fe3944fe21602fdcd2148722df48154d66bf6cb6b3f34d8"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.204592 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nfwlh" event={"ID":"4f7a80f9-c555-4325-b747-1a5a6e39e6ce","Type":"ContainerStarted","Data":"23ea02de5643673823f75650fddbcb6d953bfab3d7f27e4635e48790e0445234"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.217706 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" podStartSLOduration=123.217686926 podStartE2EDuration="2m3.217686926s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.216879506 +0000 UTC m=+141.788260775" watchObservedRunningTime="2025-12-12 06:50:54.217686926 +0000 UTC m=+141.789068195" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.222577 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" event={"ID":"2ab99367-5422-43bd-97d9-550e989d7542","Type":"ContainerStarted","Data":"0a346baba5dab5f20af62e127e099edd8715652392a9eb38685e989a3f859b64"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.262548 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nfwlh" podStartSLOduration=8.262524148 podStartE2EDuration="8.262524148s" podCreationTimestamp="2025-12-12 06:50:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.259328381 +0000 UTC m=+141.830709650" watchObservedRunningTime="2025-12-12 06:50:54.262524148 +0000 UTC m=+141.833905417" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.272051 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.274630 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.774613847 +0000 UTC m=+142.345995116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.280155 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v4nks" event={"ID":"894bf2b1-e262-41f9-94d1-b2a2e8e8b86e","Type":"ContainerStarted","Data":"5ded9e8b33dd55ff36a9103dae7d9cce67940493d1a582aa4f48ff31026b5b1d"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.299957 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6gl7z" podStartSLOduration=124.299938762 podStartE2EDuration="2m4.299938762s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.29899269 +0000 UTC m=+141.870373969" watchObservedRunningTime="2025-12-12 06:50:54.299938762 +0000 UTC m=+141.871320031" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.302322 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" event={"ID":"32c762a6-9bfc-40d6-8a78-eb4189e885bd","Type":"ContainerStarted","Data":"a8929b2dc908ead36c3325990a96c8e7e5c73b8ef8d50d2f338edb4b423b272d"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.302394 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" event={"ID":"32c762a6-9bfc-40d6-8a78-eb4189e885bd","Type":"ContainerStarted","Data":"a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.303770 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.324545 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" event={"ID":"9aab65fa-d90b-489e-8443-07cac98cb8ee","Type":"ContainerDied","Data":"f4fa4b944ca3448e70d4832e40c7302dd8efc64df35b661187f772c15f6d94a9"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.324466 4867 generic.go:334] "Generic (PLEG): container finished" podID="9aab65fa-d90b-489e-8443-07cac98cb8ee" containerID="f4fa4b944ca3448e70d4832e40c7302dd8efc64df35b661187f772c15f6d94a9" exitCode=0 Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.324670 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" event={"ID":"9aab65fa-d90b-489e-8443-07cac98cb8ee","Type":"ContainerStarted","Data":"ab4600ad893ae61588944705fe4619779fb963ecf7a62d830724c69793225962"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.327393 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" event={"ID":"37ad992d-0e36-434e-9d11-106ef955c5dd","Type":"ContainerStarted","Data":"823aaef7d3679c19382e449d2b9a5a00ad525209d66d1b1310b491103bce4f6b"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.351450 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.353759 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" event={"ID":"3d58b26a-54e5-4699-9a2c-e4af556a4405","Type":"ContainerStarted","Data":"26e9ecac3633b9edfaaf0e0b6d8e8e17a46d3df31aad6ad3a77e67cacf98eb53"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.353847 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" podStartSLOduration=124.353835381 podStartE2EDuration="2m4.353835381s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.352742135 +0000 UTC m=+141.924123404" watchObservedRunningTime="2025-12-12 06:50:54.353835381 +0000 UTC m=+141.925216650" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.357361 4867 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-45697 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.357415 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.370409 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.370634 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.376052 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.377824 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.877805504 +0000 UTC m=+142.449186773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.386164 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" event={"ID":"f8f8fec8-9201-4083-8948-6c2958ef5ec7","Type":"ContainerStarted","Data":"9b56c8d56d3ed1c611eb7adf9794fc00705e4ecb54e06bc9c8b61927059968ef"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.396417 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.403396 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" event={"ID":"fc4a337f-524f-40d0-a5d3-4f5653a2c5cd","Type":"ContainerStarted","Data":"3daa485e37d1f3d49a7344e28181b14151dd8415f4fb5f7d2b82bd1c59d34073"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.473929 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" event={"ID":"12a7c8be-2580-4a95-a465-eaac8f373729","Type":"ContainerStarted","Data":"26ed8e457033540d26afb4a88eab10b799c165138a98ef5c70cdd0e6f0c7153d"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.489658 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.492746 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:54.992727232 +0000 UTC m=+142.564108591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.499098 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" podStartSLOduration=123.499080393 podStartE2EDuration="2m3.499080393s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.488557692 +0000 UTC m=+142.059938961" watchObservedRunningTime="2025-12-12 06:50:54.499080393 +0000 UTC m=+142.070461662" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.527997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" event={"ID":"818edab4-a66c-4739-bead-9d42292e468a","Type":"ContainerStarted","Data":"620ef575bf43d00f0af40183b6f8770cfdb8d0e16c8b8fc32e65ee7733392673"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.528036 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" event={"ID":"818edab4-a66c-4739-bead-9d42292e468a","Type":"ContainerStarted","Data":"b7c3cdb90ea2d39f381ebf577dd94b1b56326e677c08c7fb1dc28788aac64bba"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.555460 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" event={"ID":"45a95b26-1418-4107-bd27-0f67e3e6b9db","Type":"ContainerStarted","Data":"2a948f67332d474f7ace2d58d20939290e550770f3ef3d0c581376866348ede6"} Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.560648 4867 patch_prober.go:28] interesting pod/downloads-7954f5f757-nmwfj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.560702 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nmwfj" podUID="a648d31a-4689-434c-b89b-c0d96065ce10" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.579631 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pjj4n" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.594103 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.595881 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.095864928 +0000 UTC m=+142.667246197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.605607 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sp4sz" podStartSLOduration=124.6055866 podStartE2EDuration="2m4.6055866s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.527711518 +0000 UTC m=+142.099092777" watchObservedRunningTime="2025-12-12 06:50:54.6055866 +0000 UTC m=+142.176967879" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.607014 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" podStartSLOduration=123.607001634 podStartE2EDuration="2m3.607001634s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.604837842 +0000 UTC m=+142.176219111" watchObservedRunningTime="2025-12-12 06:50:54.607001634 +0000 UTC m=+142.178382903" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.657285 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.663083 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" podStartSLOduration=123.663063034 podStartE2EDuration="2m3.663063034s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.660527693 +0000 UTC m=+142.231908982" watchObservedRunningTime="2025-12-12 06:50:54.663063034 +0000 UTC m=+142.234444313" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.673573 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.695837 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.696285 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.196272858 +0000 UTC m=+142.767654137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.797517 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.797970 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.297955039 +0000 UTC m=+142.869336308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.809332 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.824413 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-vkl2j" podStartSLOduration=124.824394471 podStartE2EDuration="2m4.824394471s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:54.814033703 +0000 UTC m=+142.385414972" watchObservedRunningTime="2025-12-12 06:50:54.824394471 +0000 UTC m=+142.395775740" Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.913822 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:54 crc kubenswrapper[4867]: E1212 06:50:54.914291 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.4142754 +0000 UTC m=+142.985656669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.944803 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:54 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:54 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:54 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:54 crc kubenswrapper[4867]: I1212 06:50:54.944853 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.001206 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.011721 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.016825 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.017156 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.51714141 +0000 UTC m=+143.088522679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.052626 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.063720 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.125950 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.125989 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.126037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.126098 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wf6z\" (UniqueName: \"kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.139748 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.63972931 +0000 UTC m=+143.211110579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.231205 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.231538 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wf6z\" (UniqueName: \"kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.231582 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.231610 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.232441 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.232506 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.732491568 +0000 UTC m=+143.303872827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.232990 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.272376 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wf6z\" (UniqueName: \"kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z\") pod \"redhat-marketplace-jpvvh\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.332690 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.333363 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.833350659 +0000 UTC m=+143.404731928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.338944 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.340119 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.374677 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.375934 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.433879 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.434029 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzzrx\" (UniqueName: \"kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.434056 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.434169 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.434273 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:55.934257822 +0000 UTC m=+143.505639091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.536989 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.537308 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.537409 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzzrx\" (UniqueName: \"kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.537494 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.537958 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.538299 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.03828811 +0000 UTC m=+143.609669379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.538706 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.567215 4867 generic.go:334] "Generic (PLEG): container finished" podID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerID="62f106f7bd3658a364573de0252e62305e77ac0b70eefe1cf3660a2d0d4456bf" exitCode=0 Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.567288 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerDied","Data":"62f106f7bd3658a364573de0252e62305e77ac0b70eefe1cf3660a2d0d4456bf"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.567312 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerStarted","Data":"da6d9f7f41d176f8e7f039141099bab528e2721232893ca0c7568f26d979409a"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.568805 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.583181 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" event={"ID":"81885549-e08a-4636-9f50-fb5116790c37","Type":"ContainerStarted","Data":"2495cb46aa9784e3bbf510ee8f00f6f68e113df69e8be6c5b39aab63e0cd0a4c"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.583288 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" event={"ID":"81885549-e08a-4636-9f50-fb5116790c37","Type":"ContainerStarted","Data":"40307f3dee41de60e902c54536e1b13b5fe8d6e39ae4b59c1557f419e30ea344"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.592329 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzzrx\" (UniqueName: \"kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx\") pod \"redhat-marketplace-smggn\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.615147 4867 generic.go:334] "Generic (PLEG): container finished" podID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerID="edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d" exitCode=0 Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.615502 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerDied","Data":"edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.615527 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerStarted","Data":"ef5851eef1ce8a5738abc8023ca7de12331efc797fa383e2a32a3a115d20076b"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.625032 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-njxxd" podStartSLOduration=125.625008063 podStartE2EDuration="2m5.625008063s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.623078636 +0000 UTC m=+143.194459905" watchObservedRunningTime="2025-12-12 06:50:55.625008063 +0000 UTC m=+143.196389332" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.638958 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.639178 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.139160151 +0000 UTC m=+143.710541420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.639302 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.640482 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.140473643 +0000 UTC m=+143.711854922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.664883 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" event={"ID":"2bc83e9c-4e30-45b0-aefe-bd648a8aefe4","Type":"ContainerStarted","Data":"f4e161cf103e956824f9e65ed79f00ec05c5f5e3da6f34ddd99ed17ffc9a1a53"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.664998 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.678081 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.690188 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" event={"ID":"93f987df-943c-45a9-aaf8-a0430358ba7f","Type":"ContainerStarted","Data":"23faa7041c59decada291a93579c2a5c0dd8a8b7864999664fd4609b741fabbe"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.690221 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" event={"ID":"93f987df-943c-45a9-aaf8-a0430358ba7f","Type":"ContainerStarted","Data":"5e8b615d294695d75538c01a2179a8518c60b1bfe729c8559f6740fc0638301f"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.699326 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" podStartSLOduration=124.699312329 podStartE2EDuration="2m4.699312329s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.692857345 +0000 UTC m=+143.264238614" watchObservedRunningTime="2025-12-12 06:50:55.699312329 +0000 UTC m=+143.270693598" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.712768 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-52qk7" event={"ID":"7b0920c8-ab26-4b72-b22b-c4123498a461","Type":"ContainerStarted","Data":"269e5e84e5484a596e81cd07427049de4fa74ca9cc80a22fe270c96c84678a24"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.721993 4867 generic.go:334] "Generic (PLEG): container finished" podID="739851dc-9291-46a6-bea7-0c89156eacd1" containerID="555548831c6736123f563b9cd2011d3f76f74497e6034731108fdf7f45e940de" exitCode=0 Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.722117 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerDied","Data":"555548831c6736123f563b9cd2011d3f76f74497e6034731108fdf7f45e940de"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.722145 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerStarted","Data":"fd15f3f514a883c154fda06cee0a7ad630036a769c110b1956da8b65c6e3ff54"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.728868 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-s66s9" podStartSLOduration=124.728852086 podStartE2EDuration="2m4.728852086s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.726926579 +0000 UTC m=+143.298307848" watchObservedRunningTime="2025-12-12 06:50:55.728852086 +0000 UTC m=+143.300233355" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.742065 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.743017 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.243001544 +0000 UTC m=+143.814382813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.754625 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-w6q4w" event={"ID":"a8091880-bdd4-4a8d-8078-dec177dd01bd","Type":"ContainerStarted","Data":"50326a3ba5580650f7ca2444d6cd07a7e289cc66b4df789591a8f770450f2944"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.760689 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.761938 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.781199 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.799030 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.806257 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerStarted","Data":"eaae5e77d81d72cea32b6e76537f85ff052ddc2ea64e432f6fe4c99f878eddf2"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.807391 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.808616 4867 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qlb8v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.808731 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.843457 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.843498 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.843536 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw2kz\" (UniqueName: \"kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.843579 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.845197 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.345180747 +0000 UTC m=+143.916562016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.850549 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-n64dp" event={"ID":"3d58b26a-54e5-4699-9a2c-e4af556a4405","Type":"ContainerStarted","Data":"f648c4e758fd1e2e570930fd4b8e0d6e90d30fd44886536178400d10ee851e92"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.858212 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" podStartSLOduration=124.858197848 podStartE2EDuration="2m4.858197848s" podCreationTimestamp="2025-12-12 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.85620419 +0000 UTC m=+143.427585449" watchObservedRunningTime="2025-12-12 06:50:55.858197848 +0000 UTC m=+143.429579117" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.866838 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fzxx5" event={"ID":"22c4c88c-1195-4f05-b3b4-f1308c7d5e68","Type":"ContainerStarted","Data":"e8179b5f62d0e1d13bab591d812aeec07aeaf102f130ab0b14a092b84e087e7a"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.866882 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fzxx5" event={"ID":"22c4c88c-1195-4f05-b3b4-f1308c7d5e68","Type":"ContainerStarted","Data":"6876ba3c4a7e1a88f5d03b0fde0f18ef88fb2b52f213e831b070deb2e648a057"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.879629 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" event={"ID":"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef","Type":"ContainerStarted","Data":"caab92e0a74907bb2c160eb7753354cb783b9b2bbeb8663a799b30931692830e"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.909089 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-fzxx5" podStartSLOduration=9.909074255 podStartE2EDuration="9.909074255s" podCreationTimestamp="2025-12-12 06:50:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.903591413 +0000 UTC m=+143.474972682" watchObservedRunningTime="2025-12-12 06:50:55.909074255 +0000 UTC m=+143.480455524" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.924696 4867 generic.go:334] "Generic (PLEG): container finished" podID="12a7c8be-2580-4a95-a465-eaac8f373729" containerID="c2d56966322b911c87d923113f428049314a71dea926b58cc8bd31a9774e7a8b" exitCode=0 Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.924796 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" event={"ID":"12a7c8be-2580-4a95-a465-eaac8f373729","Type":"ContainerStarted","Data":"a14d40dc9a359d1b1c7a52d20015b781680b8b6a8b4f91a154ef4f54bdd78f00"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.924822 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" event={"ID":"12a7c8be-2580-4a95-a465-eaac8f373729","Type":"ContainerDied","Data":"c2d56966322b911c87d923113f428049314a71dea926b58cc8bd31a9774e7a8b"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.942350 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:55 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:55 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:55 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.942409 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.945283 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.945549 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.945571 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.945630 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw2kz\" (UniqueName: \"kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: E1212 06:50:55.946513 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.446499369 +0000 UTC m=+144.017880638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.947154 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.947627 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.947764 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.948487 4867 generic.go:334] "Generic (PLEG): container finished" podID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerID="6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a" exitCode=0 Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.958670 4867 patch_prober.go:28] interesting pod/downloads-7954f5f757-nmwfj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" start-of-body= Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.958714 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nmwfj" podUID="a648d31a-4689-434c-b89b-c0d96065ce10" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.6:8080/\": dial tcp 10.217.0.6:8080: connect: connection refused" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.961313 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerDied","Data":"6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.961347 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerStarted","Data":"870a9a5efcce36362fb3857d261fae6a8ebe80dee9add680ca935d7ea7d4487f"} Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.961435 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.975717 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.979881 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.981861 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-d6p85" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.982195 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" podStartSLOduration=125.982183772 podStartE2EDuration="2m5.982183772s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:55.979629191 +0000 UTC m=+143.551010460" watchObservedRunningTime="2025-12-12 06:50:55.982183772 +0000 UTC m=+143.553565041" Dec 12 06:50:55 crc kubenswrapper[4867]: I1212 06:50:55.994108 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-77bzq" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:55.999340 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw2kz\" (UniqueName: \"kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz\") pod \"redhat-operators-rr6lx\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.014056 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-72qzg" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.052421 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqsh7\" (UniqueName: \"kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.062364 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.062455 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.062655 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.069057 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.069219 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.569166512 +0000 UTC m=+144.140547781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.112636 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.164781 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.165273 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqsh7\" (UniqueName: \"kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.165310 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.165333 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.175670 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.675634087 +0000 UTC m=+144.247015356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.175796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.176213 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.676205391 +0000 UTC m=+144.247586660 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.176782 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.183185 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.271083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqsh7\" (UniqueName: \"kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7\") pod \"redhat-operators-6kt9s\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.276699 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.277188 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.777171655 +0000 UTC m=+144.348552924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.352325 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.361067 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.378062 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.378498 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.878484627 +0000 UTC m=+144.449865896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: W1212 06:50:56.401869 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57754cdc_d065_4fcd_9205_78ebcddf04a7.slice/crio-d225098af6293ddfac8eb79795573bfd5e9956e9401e507cc3957b6d32f0178b WatchSource:0}: Error finding container d225098af6293ddfac8eb79795573bfd5e9956e9401e507cc3957b6d32f0178b: Status 404 returned error can't find the container with id d225098af6293ddfac8eb79795573bfd5e9956e9401e507cc3957b6d32f0178b Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.419692 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-fzxx5" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.483725 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.484116 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:56.984099442 +0000 UTC m=+144.555480711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.590964 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.591272 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.091261075 +0000 UTC m=+144.662642344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.693075 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.693734 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.193720164 +0000 UTC m=+144.765101433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.752722 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:50:56 crc kubenswrapper[4867]: W1212 06:50:56.790180 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2cda128_4d99_448e_9883_b53bae064a1b.slice/crio-3fec0ccd5b6350d8019904e832f071040a55f8556e42e18dd6d856188df85b9b WatchSource:0}: Error finding container 3fec0ccd5b6350d8019904e832f071040a55f8556e42e18dd6d856188df85b9b: Status 404 returned error can't find the container with id 3fec0ccd5b6350d8019904e832f071040a55f8556e42e18dd6d856188df85b9b Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.795483 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.795838 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.295827255 +0000 UTC m=+144.867208524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.905538 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.905721 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.405696982 +0000 UTC m=+144.977078251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.905841 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.907601 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:50:56 crc kubenswrapper[4867]: E1212 06:50:56.910409 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.410392045 +0000 UTC m=+144.981773314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.941422 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:56 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:56 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:56 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.941475 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:56 crc kubenswrapper[4867]: W1212 06:50:56.973672 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd982b2db_6664_4201_b701_aaaa17fc7bba.slice/crio-06ff0262fa18087448080d70088e3b30401b6c10b8e9cfb28d812d2982658fc4 WatchSource:0}: Error finding container 06ff0262fa18087448080d70088e3b30401b6c10b8e9cfb28d812d2982658fc4: Status 404 returned error can't find the container with id 06ff0262fa18087448080d70088e3b30401b6c10b8e9cfb28d812d2982658fc4 Dec 12 06:50:56 crc kubenswrapper[4867]: I1212 06:50:56.988515 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" event={"ID":"12a7c8be-2580-4a95-a465-eaac8f373729","Type":"ContainerStarted","Data":"cc1101b185b338c0219f76609263108d54b63bf21343f0618bbeffdacfcd5ea4"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.007757 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.007805 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" event={"ID":"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef","Type":"ContainerStarted","Data":"667c39cc26c2f5ced0ee75ab2302cce109448b433a5b5a7300550b770e400a59"} Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.007974 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.507957537 +0000 UTC m=+145.079338806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.008054 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.008354 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.508345847 +0000 UTC m=+145.079727116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.017533 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerStarted","Data":"3fec0ccd5b6350d8019904e832f071040a55f8556e42e18dd6d856188df85b9b"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.028738 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerStarted","Data":"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.028789 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerStarted","Data":"d225098af6293ddfac8eb79795573bfd5e9956e9401e507cc3957b6d32f0178b"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.051662 4867 generic.go:334] "Generic (PLEG): container finished" podID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerID="43f6b5f3ae4b7147223839366c8fcfbca7f942d4d35580b6acccb8a2e78aad76" exitCode=0 Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.052663 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerDied","Data":"43f6b5f3ae4b7147223839366c8fcfbca7f942d4d35580b6acccb8a2e78aad76"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.052686 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerStarted","Data":"d2a645b87a51fe1b9f169536da80d4716be8c486a9cd62cb30fa1806c98f0a95"} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.072958 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.095381 4867 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.112978 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.113393 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.613371838 +0000 UTC m=+145.184753107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.117249 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.125502 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.625485997 +0000 UTC m=+145.196867256 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.225469 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.225947 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.725926748 +0000 UTC m=+145.297308037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.327383 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.327731 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.827720142 +0000 UTC m=+145.399101401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.428624 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.428948 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:57.928932742 +0000 UTC m=+145.500314001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.530057 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.530438 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.030421959 +0000 UTC m=+145.601803228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.631510 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.631948 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.131932865 +0000 UTC m=+145.703314134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.734161 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.734630 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.23461354 +0000 UTC m=+145.805994809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.837556 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.838021 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.338006672 +0000 UTC m=+145.909387941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.938565 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:57 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:57 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:57 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.938618 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.939382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:57 crc kubenswrapper[4867]: E1212 06:50:57.939780 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-12 06:50:58.439767695 +0000 UTC m=+146.011148964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zcz7" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.994503 4867 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-12T06:50:57.099901385Z","Handler":null,"Name":""} Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.997621 4867 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 12 06:50:57 crc kubenswrapper[4867]: I1212 06:50:57.997660 4867 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.042857 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.120506 4867 generic.go:334] "Generic (PLEG): container finished" podID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerID="aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156" exitCode=0 Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.120644 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerDied","Data":"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.123616 4867 generic.go:334] "Generic (PLEG): container finished" podID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerID="e3eddc08304df50fac61f7b1f6112e6f46f2c52c6ada9ffc582d8c85732f83cc" exitCode=0 Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.123733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerDied","Data":"e3eddc08304df50fac61f7b1f6112e6f46f2c52c6ada9ffc582d8c85732f83cc"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.123767 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerStarted","Data":"06ff0262fa18087448080d70088e3b30401b6c10b8e9cfb28d812d2982658fc4"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.131497 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" event={"ID":"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef","Type":"ContainerStarted","Data":"f53e2e97a4f8b4571f3083f1e914408f117350f59513a38433588c7939e91738"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.131551 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" event={"ID":"b3f15d74-7f84-4f87-8b11-7ec1f993f3ef","Type":"ContainerStarted","Data":"45b3c7fe2a325337e8fbd90051411bcfdb06b1bdb3c1bc031746a9b5f48a9de1"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.134878 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.146916 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.150055 4867 generic.go:334] "Generic (PLEG): container finished" podID="f2cda128-4d99-448e-9883-b53bae064a1b" containerID="7ad07b38d4e5efd75cf22e8691d00bcae6f2da59b529c300017fcfcf5581653b" exitCode=0 Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.150155 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerDied","Data":"7ad07b38d4e5efd75cf22e8691d00bcae6f2da59b529c300017fcfcf5581653b"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.155011 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.155041 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.168025 4867 generic.go:334] "Generic (PLEG): container finished" podID="f8f8fec8-9201-4083-8948-6c2958ef5ec7" containerID="9b56c8d56d3ed1c611eb7adf9794fc00705e4ecb54e06bc9c8b61927059968ef" exitCode=0 Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.168215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" event={"ID":"f8f8fec8-9201-4083-8948-6c2958ef5ec7","Type":"ContainerDied","Data":"9b56c8d56d3ed1c611eb7adf9794fc00705e4ecb54e06bc9c8b61927059968ef"} Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.223874 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zcz7\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.235819 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t8lml" podStartSLOduration=12.235800223 podStartE2EDuration="12.235800223s" podCreationTimestamp="2025-12-12 06:50:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:50:58.233486688 +0000 UTC m=+145.804867957" watchObservedRunningTime="2025-12-12 06:50:58.235800223 +0000 UTC m=+145.807181492" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.470909 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.556069 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.569677 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.665142 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.757860 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.758311 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.758343 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.759159 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.767698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.767715 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.853621 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.929443 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.935058 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.937624 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:58 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:58 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:58 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.937680 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:50:58 crc kubenswrapper[4867]: W1212 06:50:58.952543 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-5740becfd9c4e024c8a3712439cdd40d6dd1f5227b8d40f296c72d980ab53f69 WatchSource:0}: Error finding container 5740becfd9c4e024c8a3712439cdd40d6dd1f5227b8d40f296c72d980ab53f69: Status 404 returned error can't find the container with id 5740becfd9c4e024c8a3712439cdd40d6dd1f5227b8d40f296c72d980ab53f69 Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.981990 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.988978 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:50:58 crc kubenswrapper[4867]: I1212 06:50:58.989016 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.008990 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ztz7d" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.051638 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-nmwfj" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.057335 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.094744 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.094773 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.098046 4867 patch_prober.go:28] interesting pod/console-f9d7485db-6xxwt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.098085 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6xxwt" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.187486 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5740becfd9c4e024c8a3712439cdd40d6dd1f5227b8d40f296c72d980ab53f69"} Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.193955 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" event={"ID":"4cc708f7-46a4-4947-8ba9-c960d302a348","Type":"ContainerStarted","Data":"8135db2195c7eb173f3f4a1de78c783ac531745d6ad0576d36535dd81faea807"} Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.504068 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.504723 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.507155 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.507334 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.520809 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.588448 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.588564 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.687857 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.697629 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.697680 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.697767 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: W1212 06:50:59.726814 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-545e9accd824071b7cb2bc224227055a9c790c71670c385310d16d4353b73866 WatchSource:0}: Error finding container 545e9accd824071b7cb2bc224227055a9c790c71670c385310d16d4353b73866: Status 404 returned error can't find the container with id 545e9accd824071b7cb2bc224227055a9c790c71670c385310d16d4353b73866 Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.727203 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.798137 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qn79\" (UniqueName: \"kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79\") pod \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.798197 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume\") pod \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.798236 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume\") pod \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\" (UID: \"f8f8fec8-9201-4083-8948-6c2958ef5ec7\") " Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.799278 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume" (OuterVolumeSpecName: "config-volume") pod "f8f8fec8-9201-4083-8948-6c2958ef5ec7" (UID: "f8f8fec8-9201-4083-8948-6c2958ef5ec7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.803622 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79" (OuterVolumeSpecName: "kube-api-access-5qn79") pod "f8f8fec8-9201-4083-8948-6c2958ef5ec7" (UID: "f8f8fec8-9201-4083-8948-6c2958ef5ec7"). InnerVolumeSpecName "kube-api-access-5qn79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.803777 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f8f8fec8-9201-4083-8948-6c2958ef5ec7" (UID: "f8f8fec8-9201-4083-8948-6c2958ef5ec7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.833369 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.891589 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.899537 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qn79\" (UniqueName: \"kubernetes.io/projected/f8f8fec8-9201-4083-8948-6c2958ef5ec7-kube-api-access-5qn79\") on node \"crc\" DevicePath \"\"" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.899569 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8f8fec8-9201-4083-8948-6c2958ef5ec7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.899581 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8f8fec8-9201-4083-8948-6c2958ef5ec7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.939875 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:50:59 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:50:59 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:50:59 crc kubenswrapper[4867]: healthz check failed Dec 12 06:50:59 crc kubenswrapper[4867]: I1212 06:50:59.940001 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.214981 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 12 06:51:00 crc kubenswrapper[4867]: W1212 06:51:00.224464 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc8012af1_49cc_4e27_a0f5_f7da326d57f5.slice/crio-39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d WatchSource:0}: Error finding container 39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d: Status 404 returned error can't find the container with id 39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.227673 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"005bd6bc5d0c89b35b0c1d6c5f1eb6c9e26d11eebd356c82fceb22a973209427"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.264966 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" event={"ID":"4cc708f7-46a4-4947-8ba9-c960d302a348","Type":"ContainerStarted","Data":"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.265026 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.268693 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fb852f060773fa1d2f21867dd401e8a92c446b0e115cc5d404bc63199a63f99b"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.268750 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"545e9accd824071b7cb2bc224227055a9c790c71670c385310d16d4353b73866"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.273599 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"adb1fa418c4824c8333ea51c9256b03a3fc330bc80dfc18bbd5a7d7e83451cf2"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.273631 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c2a082042ef1ea20c0afc0624422eee63d355e0bfbe3e14f98b85ae9cfeebb8b"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.274090 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.276332 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" event={"ID":"f8f8fec8-9201-4083-8948-6c2958ef5ec7","Type":"ContainerDied","Data":"9e2110d497e4563b3c8332e160c4fc5b1518fa3a31c4dcbcbbc4b73a291dec06"} Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.276360 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e2110d497e4563b3c8332e160c4fc5b1518fa3a31c4dcbcbbc4b73a291dec06" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.276413 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.307634 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" podStartSLOduration=130.307609427 podStartE2EDuration="2m10.307609427s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:51:00.283957651 +0000 UTC m=+147.855338940" watchObservedRunningTime="2025-12-12 06:51:00.307609427 +0000 UTC m=+147.878990696" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.341820 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.342568 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.348027 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.937843 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:00 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:00 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:00 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:00 crc kubenswrapper[4867]: I1212 06:51:00.937903 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.329350 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c8012af1-49cc-4e27-a0f5-f7da326d57f5","Type":"ContainerStarted","Data":"2864b11e8c2c175a89d10d4ed07ee2c68243e28e0057e9cbe2b2afda45d03d91"} Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.329674 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c8012af1-49cc-4e27-a0f5-f7da326d57f5","Type":"ContainerStarted","Data":"39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d"} Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.334866 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-c7vxj" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.349805 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.349786493 podStartE2EDuration="2.349786493s" podCreationTimestamp="2025-12-12 06:50:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:51:01.347992921 +0000 UTC m=+148.919374190" watchObservedRunningTime="2025-12-12 06:51:01.349786493 +0000 UTC m=+148.921167762" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.575070 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 12 06:51:01 crc kubenswrapper[4867]: E1212 06:51:01.575457 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f8fec8-9201-4083-8948-6c2958ef5ec7" containerName="collect-profiles" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.575475 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f8fec8-9201-4083-8948-6c2958ef5ec7" containerName="collect-profiles" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.575692 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f8fec8-9201-4083-8948-6c2958ef5ec7" containerName="collect-profiles" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.576239 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.579601 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.579887 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.594463 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.726287 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.726380 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.828320 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.828440 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.828788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.865677 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.913691 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.938791 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:01 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:01 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:01 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:01 crc kubenswrapper[4867]: I1212 06:51:01.938881 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:02 crc kubenswrapper[4867]: I1212 06:51:02.504867 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 12 06:51:02 crc kubenswrapper[4867]: W1212 06:51:02.539287 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod3dc16a46_c2ec_4769_b31c_ee9fc9e4e4f5.slice/crio-3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad WatchSource:0}: Error finding container 3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad: Status 404 returned error can't find the container with id 3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad Dec 12 06:51:02 crc kubenswrapper[4867]: I1212 06:51:02.939472 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:02 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:02 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:02 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:02 crc kubenswrapper[4867]: I1212 06:51:02.939963 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:03 crc kubenswrapper[4867]: I1212 06:51:03.350100 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5","Type":"ContainerStarted","Data":"3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad"} Dec 12 06:51:03 crc kubenswrapper[4867]: I1212 06:51:03.354644 4867 generic.go:334] "Generic (PLEG): container finished" podID="c8012af1-49cc-4e27-a0f5-f7da326d57f5" containerID="2864b11e8c2c175a89d10d4ed07ee2c68243e28e0057e9cbe2b2afda45d03d91" exitCode=0 Dec 12 06:51:03 crc kubenswrapper[4867]: I1212 06:51:03.354821 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c8012af1-49cc-4e27-a0f5-f7da326d57f5","Type":"ContainerDied","Data":"2864b11e8c2c175a89d10d4ed07ee2c68243e28e0057e9cbe2b2afda45d03d91"} Dec 12 06:51:03 crc kubenswrapper[4867]: I1212 06:51:03.942503 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:03 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:03 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:03 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:03 crc kubenswrapper[4867]: I1212 06:51:03.942573 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.627851 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.675713 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access\") pod \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.675833 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir\") pod \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\" (UID: \"c8012af1-49cc-4e27-a0f5-f7da326d57f5\") " Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.675976 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c8012af1-49cc-4e27-a0f5-f7da326d57f5" (UID: "c8012af1-49cc-4e27-a0f5-f7da326d57f5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.676215 4867 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.695464 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c8012af1-49cc-4e27-a0f5-f7da326d57f5" (UID: "c8012af1-49cc-4e27-a0f5-f7da326d57f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.777686 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c8012af1-49cc-4e27-a0f5-f7da326d57f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.938109 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:04 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:04 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:04 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:04 crc kubenswrapper[4867]: I1212 06:51:04.938373 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.375056 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c8012af1-49cc-4e27-a0f5-f7da326d57f5","Type":"ContainerDied","Data":"39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d"} Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.375104 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39b8f7c7d88582414043f068262c6e4f754c64630f55659a8fc8b4bebbaaba8d" Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.375101 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.377326 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5","Type":"ContainerStarted","Data":"137b6d283ce9f2363b858a0daec7caddd4b2f43319f332f864374f235a324b00"} Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.395955 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.395940341 podStartE2EDuration="4.395940341s" podCreationTimestamp="2025-12-12 06:51:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:51:05.38836725 +0000 UTC m=+152.959748519" watchObservedRunningTime="2025-12-12 06:51:05.395940341 +0000 UTC m=+152.967321610" Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.414538 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-fzxx5" Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.939717 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:05 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:05 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:05 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:05 crc kubenswrapper[4867]: I1212 06:51:05.940205 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:06 crc kubenswrapper[4867]: I1212 06:51:06.393048 4867 generic.go:334] "Generic (PLEG): container finished" podID="3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" containerID="137b6d283ce9f2363b858a0daec7caddd4b2f43319f332f864374f235a324b00" exitCode=0 Dec 12 06:51:06 crc kubenswrapper[4867]: I1212 06:51:06.393104 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5","Type":"ContainerDied","Data":"137b6d283ce9f2363b858a0daec7caddd4b2f43319f332f864374f235a324b00"} Dec 12 06:51:06 crc kubenswrapper[4867]: I1212 06:51:06.937820 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:06 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:06 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:06 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:06 crc kubenswrapper[4867]: I1212 06:51:06.937875 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:07 crc kubenswrapper[4867]: I1212 06:51:07.938080 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:07 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:07 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:07 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:07 crc kubenswrapper[4867]: I1212 06:51:07.938135 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:08 crc kubenswrapper[4867]: I1212 06:51:08.936940 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:08 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:08 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:08 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:08 crc kubenswrapper[4867]: I1212 06:51:08.937294 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:09 crc kubenswrapper[4867]: I1212 06:51:09.096008 4867 patch_prober.go:28] interesting pod/console-f9d7485db-6xxwt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 12 06:51:09 crc kubenswrapper[4867]: I1212 06:51:09.096064 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6xxwt" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 12 06:51:09 crc kubenswrapper[4867]: I1212 06:51:09.937753 4867 patch_prober.go:28] interesting pod/router-default-5444994796-lrj5h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 12 06:51:09 crc kubenswrapper[4867]: [-]has-synced failed: reason withheld Dec 12 06:51:09 crc kubenswrapper[4867]: [+]process-running ok Dec 12 06:51:09 crc kubenswrapper[4867]: healthz check failed Dec 12 06:51:09 crc kubenswrapper[4867]: I1212 06:51:09.937816 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lrj5h" podUID="bd04d7ab-8933-4815-97ec-7be8b2e643ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 12 06:51:10 crc kubenswrapper[4867]: I1212 06:51:10.940027 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:51:10 crc kubenswrapper[4867]: I1212 06:51:10.944642 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lrj5h" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.240774 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.280124 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access\") pod \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.280358 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir\") pod \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\" (UID: \"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5\") " Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.280720 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" (UID: "3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.285175 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" (UID: "3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.381960 4867 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.382004 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.417479 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.417468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5","Type":"ContainerDied","Data":"3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad"} Dec 12 06:51:11 crc kubenswrapper[4867]: I1212 06:51:11.417566 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ee275ae145c31b8b17c6a3d57a0afea3e3f46f4d84168a71d399a8d7061adad" Dec 12 06:51:12 crc kubenswrapper[4867]: I1212 06:51:12.595450 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:51:12 crc kubenswrapper[4867]: I1212 06:51:12.601568 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c4f347fa-32d5-44e0-bc90-6d774fa43ce0-metrics-certs\") pod \"network-metrics-daemon-hj2lv\" (UID: \"c4f347fa-32d5-44e0-bc90-6d774fa43ce0\") " pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:51:12 crc kubenswrapper[4867]: I1212 06:51:12.865591 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hj2lv" Dec 12 06:51:18 crc kubenswrapper[4867]: I1212 06:51:18.475173 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:51:19 crc kubenswrapper[4867]: I1212 06:51:19.100635 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:51:19 crc kubenswrapper[4867]: I1212 06:51:19.105614 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 06:51:28 crc kubenswrapper[4867]: I1212 06:51:28.988570 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:51:28 crc kubenswrapper[4867]: I1212 06:51:28.989494 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:51:29 crc kubenswrapper[4867]: E1212 06:51:29.989064 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 12 06:51:29 crc kubenswrapper[4867]: E1212 06:51:29.989285 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xqsh7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6kt9s_openshift-marketplace(d982b2db-6664-4201-b701-aaaa17fc7bba): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:29 crc kubenswrapper[4867]: E1212 06:51:29.990485 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6kt9s" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" Dec 12 06:51:30 crc kubenswrapper[4867]: I1212 06:51:30.366660 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ngkbp" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.523624 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6kt9s" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.599028 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.599307 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrztq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-ncppm_openshift-marketplace(4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.600491 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-ncppm" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.655363 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.655523 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gw2kz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rr6lx_openshift-marketplace(f2cda128-4d99-448e-9883-b53bae064a1b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:35 crc kubenswrapper[4867]: E1212 06:51:35.656966 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rr6lx" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.626872 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.627338 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gzzrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-smggn_openshift-marketplace(57754cdc-d065-4fcd-9205-78ebcddf04a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.628508 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-smggn" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.641150 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.641324 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8zg4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-g678h_openshift-marketplace(b3613595-6ad9-482c-a9fd-1a3e2c5cc430): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:36 crc kubenswrapper[4867]: E1212 06:51:36.642510 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-g678h" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.172457 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.172682 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8012af1-49cc-4e27-a0f5-f7da326d57f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.172693 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8012af1-49cc-4e27-a0f5-f7da326d57f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.172707 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.172713 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.178573 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc16a46-c2ec-4769-b31c-ee9fc9e4e4f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.178610 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8012af1-49cc-4e27-a0f5-f7da326d57f5" containerName="pruner" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.178955 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.179030 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.183684 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.184440 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.319352 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.319468 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.420418 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.420503 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.420584 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.438257 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: I1212 06:51:37.531887 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.903523 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-ncppm" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.903789 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-g678h" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.903815 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-smggn" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.904530 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rr6lx" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.980828 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.981181 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7wf6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jpvvh_openshift-marketplace(8e0f2043-4b40-47c5-96ec-0e13dafcf1fa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.982289 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jpvvh" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.985610 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.985714 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-54nv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fmxsr_openshift-marketplace(739851dc-9291-46a6-bea7-0c89156eacd1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:37 crc kubenswrapper[4867]: E1212 06:51:37.987707 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fmxsr" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.017670 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.017829 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gml9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rkrfw_openshift-marketplace(3922f7cb-230c-49ce-b81a-cb23e3dc1c2d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.019050 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rkrfw" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" Dec 12 06:51:38 crc kubenswrapper[4867]: I1212 06:51:38.377861 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hj2lv"] Dec 12 06:51:38 crc kubenswrapper[4867]: I1212 06:51:38.378205 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 12 06:51:38 crc kubenswrapper[4867]: I1212 06:51:38.566736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fcdefd5b-54d5-4d04-a611-c63adedf727b","Type":"ContainerStarted","Data":"a56e85fd71ead269b372f8fb9fe8b6ccf4791eb185693f6c0890065670220938"} Dec 12 06:51:38 crc kubenswrapper[4867]: I1212 06:51:38.569002 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" event={"ID":"c4f347fa-32d5-44e0-bc90-6d774fa43ce0","Type":"ContainerStarted","Data":"ee43011a5c9a7987318b276669d57215b27488d4dd686c9399b1f265f89f158b"} Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.569778 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fmxsr" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.570142 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rkrfw" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" Dec 12 06:51:38 crc kubenswrapper[4867]: E1212 06:51:38.570949 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jpvvh" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" Dec 12 06:51:38 crc kubenswrapper[4867]: I1212 06:51:38.985514 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 12 06:51:39 crc kubenswrapper[4867]: I1212 06:51:39.574066 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" event={"ID":"c4f347fa-32d5-44e0-bc90-6d774fa43ce0","Type":"ContainerStarted","Data":"f8246a0c41fd2c842f9a2da31802f05127199916d192862fdf8161dafbc8110b"} Dec 12 06:51:39 crc kubenswrapper[4867]: I1212 06:51:39.574468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hj2lv" event={"ID":"c4f347fa-32d5-44e0-bc90-6d774fa43ce0","Type":"ContainerStarted","Data":"e81b334bc28baba544bf968c8ab6aed720060a837712d4fcb31821f7ffa37929"} Dec 12 06:51:39 crc kubenswrapper[4867]: I1212 06:51:39.575589 4867 generic.go:334] "Generic (PLEG): container finished" podID="fcdefd5b-54d5-4d04-a611-c63adedf727b" containerID="d8f838e0d0252bd87b07f15591445e66635dbe8ec0dae54102736b0ccf78201a" exitCode=0 Dec 12 06:51:39 crc kubenswrapper[4867]: I1212 06:51:39.575613 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fcdefd5b-54d5-4d04-a611-c63adedf727b","Type":"ContainerDied","Data":"d8f838e0d0252bd87b07f15591445e66635dbe8ec0dae54102736b0ccf78201a"} Dec 12 06:51:39 crc kubenswrapper[4867]: I1212 06:51:39.600078 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hj2lv" podStartSLOduration=169.600063873 podStartE2EDuration="2m49.600063873s" podCreationTimestamp="2025-12-12 06:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:51:39.599583892 +0000 UTC m=+187.170965161" watchObservedRunningTime="2025-12-12 06:51:39.600063873 +0000 UTC m=+187.171445142" Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.897056 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.960247 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access\") pod \"fcdefd5b-54d5-4d04-a611-c63adedf727b\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.960310 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir\") pod \"fcdefd5b-54d5-4d04-a611-c63adedf727b\" (UID: \"fcdefd5b-54d5-4d04-a611-c63adedf727b\") " Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.960469 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fcdefd5b-54d5-4d04-a611-c63adedf727b" (UID: "fcdefd5b-54d5-4d04-a611-c63adedf727b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.962530 4867 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fcdefd5b-54d5-4d04-a611-c63adedf727b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:40 crc kubenswrapper[4867]: I1212 06:51:40.966376 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fcdefd5b-54d5-4d04-a611-c63adedf727b" (UID: "fcdefd5b-54d5-4d04-a611-c63adedf727b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:51:41 crc kubenswrapper[4867]: I1212 06:51:41.063415 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fcdefd5b-54d5-4d04-a611-c63adedf727b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:51:41 crc kubenswrapper[4867]: I1212 06:51:41.588680 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"fcdefd5b-54d5-4d04-a611-c63adedf727b","Type":"ContainerDied","Data":"a56e85fd71ead269b372f8fb9fe8b6ccf4791eb185693f6c0890065670220938"} Dec 12 06:51:41 crc kubenswrapper[4867]: I1212 06:51:41.589045 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a56e85fd71ead269b372f8fb9fe8b6ccf4791eb185693f6c0890065670220938" Dec 12 06:51:41 crc kubenswrapper[4867]: I1212 06:51:41.588743 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.971959 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 12 06:51:43 crc kubenswrapper[4867]: E1212 06:51:43.972182 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdefd5b-54d5-4d04-a611-c63adedf727b" containerName="pruner" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.972205 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdefd5b-54d5-4d04-a611-c63adedf727b" containerName="pruner" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.972336 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdefd5b-54d5-4d04-a611-c63adedf727b" containerName="pruner" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.972683 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.975079 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.975396 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 12 06:51:43 crc kubenswrapper[4867]: I1212 06:51:43.980364 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.103356 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.103420 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.103499 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.204286 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.204358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.204392 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.204481 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.204763 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.225075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access\") pod \"installer-9-crc\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.330060 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:51:44 crc kubenswrapper[4867]: I1212 06:51:44.709011 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 12 06:51:44 crc kubenswrapper[4867]: W1212 06:51:44.713387 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda8d05d49_5e34_414b_b449_08464cf7d5be.slice/crio-64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9 WatchSource:0}: Error finding container 64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9: Status 404 returned error can't find the container with id 64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9 Dec 12 06:51:45 crc kubenswrapper[4867]: I1212 06:51:45.607412 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a8d05d49-5e34-414b-b449-08464cf7d5be","Type":"ContainerStarted","Data":"6f7cfedf70229812a59a62dbe096cfa62c008c306798b0f85087d7cb1fe6fadc"} Dec 12 06:51:45 crc kubenswrapper[4867]: I1212 06:51:45.607463 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a8d05d49-5e34-414b-b449-08464cf7d5be","Type":"ContainerStarted","Data":"64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9"} Dec 12 06:51:45 crc kubenswrapper[4867]: I1212 06:51:45.619144 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.619126951 podStartE2EDuration="2.619126951s" podCreationTimestamp="2025-12-12 06:51:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:51:45.618811043 +0000 UTC m=+193.190192312" watchObservedRunningTime="2025-12-12 06:51:45.619126951 +0000 UTC m=+193.190508220" Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.810333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerStarted","Data":"fe3c2d97e73803e5d58d8b4befcbb6fb3396771206c5d7cc129d2c5b316ec03b"} Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.813618 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerStarted","Data":"1d7271931eb38873a9afd1afbbc7c313e057214c49c55d78d4221db8dd089e56"} Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.816337 4867 generic.go:334] "Generic (PLEG): container finished" podID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerID="d91e09ed0e36cc8d98cc648413a0d2c280df8cab6684c6b0c750f0fa3d2db814" exitCode=0 Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.816456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerDied","Data":"d91e09ed0e36cc8d98cc648413a0d2c280df8cab6684c6b0c750f0fa3d2db814"} Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.820078 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerStarted","Data":"920eba8ff0d467ce4d913242776412fc84615ec43640d916f23be883f0a7eefc"} Dec 12 06:51:52 crc kubenswrapper[4867]: I1212 06:51:52.824148 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerStarted","Data":"93ea830534342a60c996bc5475ee6f3e9279ce1bc534f183cbe6ae35eba36a06"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.830765 4867 generic.go:334] "Generic (PLEG): container finished" podID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerID="fe3c2d97e73803e5d58d8b4befcbb6fb3396771206c5d7cc129d2c5b316ec03b" exitCode=0 Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.831161 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerDied","Data":"fe3c2d97e73803e5d58d8b4befcbb6fb3396771206c5d7cc129d2c5b316ec03b"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.834400 4867 generic.go:334] "Generic (PLEG): container finished" podID="f2cda128-4d99-448e-9883-b53bae064a1b" containerID="1d7271931eb38873a9afd1afbbc7c313e057214c49c55d78d4221db8dd089e56" exitCode=0 Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.834477 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerDied","Data":"1d7271931eb38873a9afd1afbbc7c313e057214c49c55d78d4221db8dd089e56"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.838249 4867 generic.go:334] "Generic (PLEG): container finished" podID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerID="04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0" exitCode=0 Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.838332 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerDied","Data":"04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.844163 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerStarted","Data":"68d3194db8704b335d0e73715c75fd1171f0020a2e181511f28802454a53819f"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.850073 4867 generic.go:334] "Generic (PLEG): container finished" podID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerID="920eba8ff0d467ce4d913242776412fc84615ec43640d916f23be883f0a7eefc" exitCode=0 Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.850163 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerDied","Data":"920eba8ff0d467ce4d913242776412fc84615ec43640d916f23be883f0a7eefc"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.856808 4867 generic.go:334] "Generic (PLEG): container finished" podID="739851dc-9291-46a6-bea7-0c89156eacd1" containerID="93ea830534342a60c996bc5475ee6f3e9279ce1bc534f183cbe6ae35eba36a06" exitCode=0 Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.856847 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerDied","Data":"93ea830534342a60c996bc5475ee6f3e9279ce1bc534f183cbe6ae35eba36a06"} Dec 12 06:51:53 crc kubenswrapper[4867]: I1212 06:51:53.875287 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpvvh" podStartSLOduration=3.6403615719999998 podStartE2EDuration="59.875271112s" podCreationTimestamp="2025-12-12 06:50:54 +0000 UTC" firstStartedPulling="2025-12-12 06:50:57.072308005 +0000 UTC m=+144.643689274" lastFinishedPulling="2025-12-12 06:51:53.307217545 +0000 UTC m=+200.878598814" observedRunningTime="2025-12-12 06:51:53.867614714 +0000 UTC m=+201.438995983" watchObservedRunningTime="2025-12-12 06:51:53.875271112 +0000 UTC m=+201.446652381" Dec 12 06:51:54 crc kubenswrapper[4867]: I1212 06:51:54.862776 4867 generic.go:334] "Generic (PLEG): container finished" podID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerID="b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0" exitCode=0 Dec 12 06:51:54 crc kubenswrapper[4867]: I1212 06:51:54.862876 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerDied","Data":"b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0"} Dec 12 06:51:54 crc kubenswrapper[4867]: I1212 06:51:54.866149 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerStarted","Data":"5efa8a3b8f05e164105cab5dd8c2c390ecc38e7a346d6908a2b656444368d1b3"} Dec 12 06:51:54 crc kubenswrapper[4867]: I1212 06:51:54.899582 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6kt9s" podStartSLOduration=3.65227799 podStartE2EDuration="59.899561735s" podCreationTimestamp="2025-12-12 06:50:55 +0000 UTC" firstStartedPulling="2025-12-12 06:50:58.135834863 +0000 UTC m=+145.707216132" lastFinishedPulling="2025-12-12 06:51:54.383118608 +0000 UTC m=+201.954499877" observedRunningTime="2025-12-12 06:51:54.897198575 +0000 UTC m=+202.468579854" watchObservedRunningTime="2025-12-12 06:51:54.899561735 +0000 UTC m=+202.470943004" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.376336 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.376407 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.440045 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.872785 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerStarted","Data":"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d"} Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.874755 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerStarted","Data":"5af7887ed4718c2c1df5b64c9bca2c9d5bb5a905f315e01c60b8dca0437dd987"} Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.876142 4867 generic.go:334] "Generic (PLEG): container finished" podID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerID="a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274" exitCode=0 Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.876199 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerDied","Data":"a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274"} Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.878511 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerStarted","Data":"443f97c96d0ae47ac5bc0e8273d01be0309b08be5c029d10795a95a26bf06d0b"} Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.880571 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerStarted","Data":"a18d4983b7ed6705e7087748f849efa32e6913d48ec963fc9ce35ef67e7bdb2a"} Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.892344 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-smggn" podStartSLOduration=3.382027883 podStartE2EDuration="1m0.892330896s" podCreationTimestamp="2025-12-12 06:50:55 +0000 UTC" firstStartedPulling="2025-12-12 06:50:57.050907614 +0000 UTC m=+144.622288883" lastFinishedPulling="2025-12-12 06:51:54.561210627 +0000 UTC m=+202.132591896" observedRunningTime="2025-12-12 06:51:55.889029911 +0000 UTC m=+203.460411180" watchObservedRunningTime="2025-12-12 06:51:55.892330896 +0000 UTC m=+203.463712165" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.915773 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rr6lx" podStartSLOduration=4.340515834 podStartE2EDuration="1m0.91575269s" podCreationTimestamp="2025-12-12 06:50:55 +0000 UTC" firstStartedPulling="2025-12-12 06:50:58.152023219 +0000 UTC m=+145.723404488" lastFinishedPulling="2025-12-12 06:51:54.727260075 +0000 UTC m=+202.298641344" observedRunningTime="2025-12-12 06:51:55.906719547 +0000 UTC m=+203.478100816" watchObservedRunningTime="2025-12-12 06:51:55.91575269 +0000 UTC m=+203.487133959" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.931789 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g678h" podStartSLOduration=4.566288022 podStartE2EDuration="1m3.931770663s" podCreationTimestamp="2025-12-12 06:50:52 +0000 UTC" firstStartedPulling="2025-12-12 06:50:55.568572553 +0000 UTC m=+143.139953822" lastFinishedPulling="2025-12-12 06:51:54.934055194 +0000 UTC m=+202.505436463" observedRunningTime="2025-12-12 06:51:55.927696688 +0000 UTC m=+203.499077957" watchObservedRunningTime="2025-12-12 06:51:55.931770663 +0000 UTC m=+203.503151932" Dec 12 06:51:55 crc kubenswrapper[4867]: I1212 06:51:55.945299 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fmxsr" podStartSLOduration=5.116825218 podStartE2EDuration="1m3.94528348s" podCreationTimestamp="2025-12-12 06:50:52 +0000 UTC" firstStartedPulling="2025-12-12 06:50:55.734663124 +0000 UTC m=+143.306044393" lastFinishedPulling="2025-12-12 06:51:54.563121386 +0000 UTC m=+202.134502655" observedRunningTime="2025-12-12 06:51:55.943836803 +0000 UTC m=+203.515218072" watchObservedRunningTime="2025-12-12 06:51:55.94528348 +0000 UTC m=+203.516664749" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.113819 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.113879 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.354359 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.354408 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.889218 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerStarted","Data":"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be"} Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.891078 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerStarted","Data":"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a"} Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.908861 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ncppm" podStartSLOduration=4.758662868 podStartE2EDuration="1m4.908839389s" podCreationTimestamp="2025-12-12 06:50:52 +0000 UTC" firstStartedPulling="2025-12-12 06:50:55.617559204 +0000 UTC m=+143.188940473" lastFinishedPulling="2025-12-12 06:51:55.767735725 +0000 UTC m=+203.339116994" observedRunningTime="2025-12-12 06:51:56.907437452 +0000 UTC m=+204.478818731" watchObservedRunningTime="2025-12-12 06:51:56.908839389 +0000 UTC m=+204.480220658" Dec 12 06:51:56 crc kubenswrapper[4867]: I1212 06:51:56.936798 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rkrfw" podStartSLOduration=3.512377196 podStartE2EDuration="1m3.936779739s" podCreationTimestamp="2025-12-12 06:50:53 +0000 UTC" firstStartedPulling="2025-12-12 06:50:55.959779477 +0000 UTC m=+143.531160746" lastFinishedPulling="2025-12-12 06:51:56.38418202 +0000 UTC m=+203.955563289" observedRunningTime="2025-12-12 06:51:56.935699141 +0000 UTC m=+204.507080420" watchObservedRunningTime="2025-12-12 06:51:56.936779739 +0000 UTC m=+204.508161008" Dec 12 06:51:57 crc kubenswrapper[4867]: I1212 06:51:57.156258 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rr6lx" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="registry-server" probeResult="failure" output=< Dec 12 06:51:57 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 06:51:57 crc kubenswrapper[4867]: > Dec 12 06:51:57 crc kubenswrapper[4867]: I1212 06:51:57.405765 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6kt9s" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="registry-server" probeResult="failure" output=< Dec 12 06:51:57 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 06:51:57 crc kubenswrapper[4867]: > Dec 12 06:51:58 crc kubenswrapper[4867]: I1212 06:51:58.988760 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:51:58 crc kubenswrapper[4867]: I1212 06:51:58.989097 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:51:58 crc kubenswrapper[4867]: I1212 06:51:58.989136 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:51:58 crc kubenswrapper[4867]: I1212 06:51:58.989655 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 06:51:58 crc kubenswrapper[4867]: I1212 06:51:58.989776 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af" gracePeriod=600 Dec 12 06:52:01 crc kubenswrapper[4867]: I1212 06:52:01.916629 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af" exitCode=0 Dec 12 06:52:01 crc kubenswrapper[4867]: I1212 06:52:01.916733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af"} Dec 12 06:52:02 crc kubenswrapper[4867]: I1212 06:52:02.923736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88"} Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.024317 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.025482 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.065913 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.312600 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.312664 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.351267 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.418315 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.418366 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.454987 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.470130 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.470177 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.504586 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.966514 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.966728 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.967154 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:03 crc kubenswrapper[4867]: I1212 06:52:03.969808 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:52:04 crc kubenswrapper[4867]: I1212 06:52:04.899759 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.428029 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.678633 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.679483 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.712708 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.897117 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.938861 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ncppm" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="registry-server" containerID="cri-o://248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be" gracePeriod=2 Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.939198 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rkrfw" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="registry-server" containerID="cri-o://33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a" gracePeriod=2 Dec 12 06:52:05 crc kubenswrapper[4867]: I1212 06:52:05.983527 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.150320 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.205106 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.344268 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.350450 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.389467 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.414604 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gml9m\" (UniqueName: \"kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m\") pod \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.414724 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities\") pod \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.414807 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content\") pod \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\" (UID: \"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.415668 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities" (OuterVolumeSpecName: "utilities") pod "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" (UID: "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.419823 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m" (OuterVolumeSpecName: "kube-api-access-gml9m") pod "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" (UID: "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d"). InnerVolumeSpecName "kube-api-access-gml9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.424861 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.515830 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities\") pod \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.515944 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrztq\" (UniqueName: \"kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq\") pod \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.516005 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content\") pod \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\" (UID: \"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00\") " Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.516205 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gml9m\" (UniqueName: \"kubernetes.io/projected/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-kube-api-access-gml9m\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.516217 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.516573 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities" (OuterVolumeSpecName: "utilities") pod "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" (UID: "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.518819 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq" (OuterVolumeSpecName: "kube-api-access-wrztq") pod "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" (UID: "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00"). InnerVolumeSpecName "kube-api-access-wrztq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.617641 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrztq\" (UniqueName: \"kubernetes.io/projected/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-kube-api-access-wrztq\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.617670 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.638721 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" (UID: "4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.719264 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.794451 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" (UID: "3922f7cb-230c-49ce-b81a-cb23e3dc1c2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.820047 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.945948 4867 generic.go:334] "Generic (PLEG): container finished" podID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerID="33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a" exitCode=0 Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.946022 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rkrfw" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.945996 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerDied","Data":"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a"} Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.946189 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rkrfw" event={"ID":"3922f7cb-230c-49ce-b81a-cb23e3dc1c2d","Type":"ContainerDied","Data":"870a9a5efcce36362fb3857d261fae6a8ebe80dee9add680ca935d7ea7d4487f"} Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.946233 4867 scope.go:117] "RemoveContainer" containerID="33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.950441 4867 generic.go:334] "Generic (PLEG): container finished" podID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerID="248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be" exitCode=0 Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.950566 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerDied","Data":"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be"} Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.950613 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncppm" event={"ID":"4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00","Type":"ContainerDied","Data":"ef5851eef1ce8a5738abc8023ca7de12331efc797fa383e2a32a3a115d20076b"} Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.950850 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncppm" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.967171 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.971603 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rkrfw"] Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.975988 4867 scope.go:117] "RemoveContainer" containerID="a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274" Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.983571 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:52:06 crc kubenswrapper[4867]: I1212 06:52:06.989845 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ncppm"] Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.015677 4867 scope.go:117] "RemoveContainer" containerID="6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.038682 4867 scope.go:117] "RemoveContainer" containerID="33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.039136 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a\": container with ID starting with 33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a not found: ID does not exist" containerID="33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.039167 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a"} err="failed to get container status \"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a\": rpc error: code = NotFound desc = could not find container \"33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a\": container with ID starting with 33e22498c873fb56c6083e810d37c86bc52ca9534f398233da5fb5136e2c028a not found: ID does not exist" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.039193 4867 scope.go:117] "RemoveContainer" containerID="a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.039491 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274\": container with ID starting with a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274 not found: ID does not exist" containerID="a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.039530 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274"} err="failed to get container status \"a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274\": rpc error: code = NotFound desc = could not find container \"a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274\": container with ID starting with a3f3e18c4d7d396bde8cebcf45cb82a220eebbacd2896da35712a9e3c77fd274 not found: ID does not exist" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.039559 4867 scope.go:117] "RemoveContainer" containerID="6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.039939 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a\": container with ID starting with 6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a not found: ID does not exist" containerID="6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.040119 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a"} err="failed to get container status \"6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a\": rpc error: code = NotFound desc = could not find container \"6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a\": container with ID starting with 6d1f28d6afa080a895f6f0bd9bd3bfc51ebf72e888a43492c992fd45a865fb6a not found: ID does not exist" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.040266 4867 scope.go:117] "RemoveContainer" containerID="248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.054864 4867 scope.go:117] "RemoveContainer" containerID="b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.083205 4867 scope.go:117] "RemoveContainer" containerID="edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.103309 4867 scope.go:117] "RemoveContainer" containerID="248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.103703 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be\": container with ID starting with 248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be not found: ID does not exist" containerID="248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.103738 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be"} err="failed to get container status \"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be\": rpc error: code = NotFound desc = could not find container \"248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be\": container with ID starting with 248b44ad7ea3320491b54ed0259ffd22214b020efd5eacac3b22a16d4962e0be not found: ID does not exist" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.103763 4867 scope.go:117] "RemoveContainer" containerID="b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.104065 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0\": container with ID starting with b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0 not found: ID does not exist" containerID="b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.104090 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0"} err="failed to get container status \"b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0\": rpc error: code = NotFound desc = could not find container \"b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0\": container with ID starting with b5d11cd993318f47650c4b5eab4198367d6982560f7817f0ebd410dc23f5afc0 not found: ID does not exist" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.104104 4867 scope.go:117] "RemoveContainer" containerID="edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d" Dec 12 06:52:07 crc kubenswrapper[4867]: E1212 06:52:07.104348 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d\": container with ID starting with edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d not found: ID does not exist" containerID="edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d" Dec 12 06:52:07 crc kubenswrapper[4867]: I1212 06:52:07.104380 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d"} err="failed to get container status \"edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d\": rpc error: code = NotFound desc = could not find container \"edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d\": container with ID starting with edaf92f9e153c387daac3f3ef02824eb999098c7db023baea16015708dcd8c8d not found: ID does not exist" Dec 12 06:52:08 crc kubenswrapper[4867]: I1212 06:52:08.293302 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:52:08 crc kubenswrapper[4867]: I1212 06:52:08.846793 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" path="/var/lib/kubelet/pods/3922f7cb-230c-49ce-b81a-cb23e3dc1c2d/volumes" Dec 12 06:52:08 crc kubenswrapper[4867]: I1212 06:52:08.847884 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" path="/var/lib/kubelet/pods/4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00/volumes" Dec 12 06:52:08 crc kubenswrapper[4867]: I1212 06:52:08.963410 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-smggn" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="registry-server" containerID="cri-o://335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d" gracePeriod=2 Dec 12 06:52:09 crc kubenswrapper[4867]: I1212 06:52:09.991647 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfk5k"] Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.553020 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.668515 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities\") pod \"57754cdc-d065-4fcd-9205-78ebcddf04a7\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.668621 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content\") pod \"57754cdc-d065-4fcd-9205-78ebcddf04a7\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.668691 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzzrx\" (UniqueName: \"kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx\") pod \"57754cdc-d065-4fcd-9205-78ebcddf04a7\" (UID: \"57754cdc-d065-4fcd-9205-78ebcddf04a7\") " Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.670279 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities" (OuterVolumeSpecName: "utilities") pod "57754cdc-d065-4fcd-9205-78ebcddf04a7" (UID: "57754cdc-d065-4fcd-9205-78ebcddf04a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.674148 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx" (OuterVolumeSpecName: "kube-api-access-gzzrx") pod "57754cdc-d065-4fcd-9205-78ebcddf04a7" (UID: "57754cdc-d065-4fcd-9205-78ebcddf04a7"). InnerVolumeSpecName "kube-api-access-gzzrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.688515 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57754cdc-d065-4fcd-9205-78ebcddf04a7" (UID: "57754cdc-d065-4fcd-9205-78ebcddf04a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.696187 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.696427 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6kt9s" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="registry-server" containerID="cri-o://5efa8a3b8f05e164105cab5dd8c2c390ecc38e7a346d6908a2b656444368d1b3" gracePeriod=2 Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.769629 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzzrx\" (UniqueName: \"kubernetes.io/projected/57754cdc-d065-4fcd-9205-78ebcddf04a7-kube-api-access-gzzrx\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.769669 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.769680 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57754cdc-d065-4fcd-9205-78ebcddf04a7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.977325 4867 generic.go:334] "Generic (PLEG): container finished" podID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerID="335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d" exitCode=0 Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.977400 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-smggn" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.977419 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerDied","Data":"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d"} Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.978244 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-smggn" event={"ID":"57754cdc-d065-4fcd-9205-78ebcddf04a7","Type":"ContainerDied","Data":"d225098af6293ddfac8eb79795573bfd5e9956e9401e507cc3957b6d32f0178b"} Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.978268 4867 scope.go:117] "RemoveContainer" containerID="335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d" Dec 12 06:52:10 crc kubenswrapper[4867]: I1212 06:52:10.999372 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.001376 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-smggn"] Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.203474 4867 scope.go:117] "RemoveContainer" containerID="04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.221947 4867 scope.go:117] "RemoveContainer" containerID="aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.240477 4867 scope.go:117] "RemoveContainer" containerID="335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d" Dec 12 06:52:11 crc kubenswrapper[4867]: E1212 06:52:11.240940 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d\": container with ID starting with 335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d not found: ID does not exist" containerID="335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.240986 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d"} err="failed to get container status \"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d\": rpc error: code = NotFound desc = could not find container \"335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d\": container with ID starting with 335cc87b5249af420f6d988921ef46a1e4dcff03baf0207cade6efd1203cf78d not found: ID does not exist" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.241013 4867 scope.go:117] "RemoveContainer" containerID="04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0" Dec 12 06:52:11 crc kubenswrapper[4867]: E1212 06:52:11.241650 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0\": container with ID starting with 04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0 not found: ID does not exist" containerID="04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.241689 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0"} err="failed to get container status \"04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0\": rpc error: code = NotFound desc = could not find container \"04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0\": container with ID starting with 04d1db1c815a64e791b5d57509e6669d9b3dcb74c60db2ab38399ad332b1d6a0 not found: ID does not exist" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.241721 4867 scope.go:117] "RemoveContainer" containerID="aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156" Dec 12 06:52:11 crc kubenswrapper[4867]: E1212 06:52:11.241937 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156\": container with ID starting with aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156 not found: ID does not exist" containerID="aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.241969 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156"} err="failed to get container status \"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156\": rpc error: code = NotFound desc = could not find container \"aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156\": container with ID starting with aab1323da03a89c40c71de940daa512c9d7a501c104d1da3d087f714c41d6156 not found: ID does not exist" Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.992539 4867 generic.go:334] "Generic (PLEG): container finished" podID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerID="5efa8a3b8f05e164105cab5dd8c2c390ecc38e7a346d6908a2b656444368d1b3" exitCode=0 Dec 12 06:52:11 crc kubenswrapper[4867]: I1212 06:52:11.992622 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerDied","Data":"5efa8a3b8f05e164105cab5dd8c2c390ecc38e7a346d6908a2b656444368d1b3"} Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.035661 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.184280 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities\") pod \"d982b2db-6664-4201-b701-aaaa17fc7bba\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.184335 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content\") pod \"d982b2db-6664-4201-b701-aaaa17fc7bba\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.184381 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqsh7\" (UniqueName: \"kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7\") pod \"d982b2db-6664-4201-b701-aaaa17fc7bba\" (UID: \"d982b2db-6664-4201-b701-aaaa17fc7bba\") " Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.185673 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities" (OuterVolumeSpecName: "utilities") pod "d982b2db-6664-4201-b701-aaaa17fc7bba" (UID: "d982b2db-6664-4201-b701-aaaa17fc7bba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.189394 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7" (OuterVolumeSpecName: "kube-api-access-xqsh7") pod "d982b2db-6664-4201-b701-aaaa17fc7bba" (UID: "d982b2db-6664-4201-b701-aaaa17fc7bba"). InnerVolumeSpecName "kube-api-access-xqsh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.287314 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.287384 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqsh7\" (UniqueName: \"kubernetes.io/projected/d982b2db-6664-4201-b701-aaaa17fc7bba-kube-api-access-xqsh7\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.307604 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d982b2db-6664-4201-b701-aaaa17fc7bba" (UID: "d982b2db-6664-4201-b701-aaaa17fc7bba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.388608 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d982b2db-6664-4201-b701-aaaa17fc7bba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:12 crc kubenswrapper[4867]: I1212 06:52:12.845748 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" path="/var/lib/kubelet/pods/57754cdc-d065-4fcd-9205-78ebcddf04a7/volumes" Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.000636 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6kt9s" event={"ID":"d982b2db-6664-4201-b701-aaaa17fc7bba","Type":"ContainerDied","Data":"06ff0262fa18087448080d70088e3b30401b6c10b8e9cfb28d812d2982658fc4"} Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.000675 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6kt9s" Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.000681 4867 scope.go:117] "RemoveContainer" containerID="5efa8a3b8f05e164105cab5dd8c2c390ecc38e7a346d6908a2b656444368d1b3" Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.015510 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.016814 4867 scope.go:117] "RemoveContainer" containerID="920eba8ff0d467ce4d913242776412fc84615ec43640d916f23be883f0a7eefc" Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.020334 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6kt9s"] Dec 12 06:52:13 crc kubenswrapper[4867]: I1212 06:52:13.039665 4867 scope.go:117] "RemoveContainer" containerID="e3eddc08304df50fac61f7b1f6112e6f46f2c52c6ada9ffc582d8c85732f83cc" Dec 12 06:52:14 crc kubenswrapper[4867]: I1212 06:52:14.844313 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" path="/var/lib/kubelet/pods/d982b2db-6664-4201-b701-aaaa17fc7bba/volumes" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.105259 4867 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106119 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106135 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106155 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106163 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106176 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106184 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106195 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106203 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106305 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106320 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106333 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106341 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106378 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106387 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106399 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106407 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106415 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106424 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106435 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106442 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="extract-content" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106458 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106468 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="extract-utilities" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.106483 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106492 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106637 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3922f7cb-230c-49ce-b81a-cb23e3dc1c2d" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106659 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d982b2db-6664-4201-b701-aaaa17fc7bba" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106671 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dfc1e7b-da03-45d0-9cfb-5e0cf488dc00" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.106707 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="57754cdc-d065-4fcd-9205-78ebcddf04a7" containerName="registry-server" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107169 4867 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107428 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107568 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817" gracePeriod=15 Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107633 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f" gracePeriod=15 Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107690 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24" gracePeriod=15 Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107717 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3" gracePeriod=15 Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107794 4867 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.107689 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2" gracePeriod=15 Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108057 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108071 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108089 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108096 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108106 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108113 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108125 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108133 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108143 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108151 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108165 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108172 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108315 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108326 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108336 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108345 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108357 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 12 06:52:23 crc kubenswrapper[4867]: E1212 06:52:23.108484 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108494 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.108603 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.112977 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.156055 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217277 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217594 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217684 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217784 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.217882 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319493 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319563 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319598 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319630 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319632 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319701 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319653 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319732 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319758 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319823 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319875 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319905 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.319911 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:23 crc kubenswrapper[4867]: I1212 06:52:23.448662 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:52:24 crc kubenswrapper[4867]: I1212 06:52:24.059374 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"da269f095355805d35629615493a8c2f9c1a2ad4c2bcf1ef366ca508976e4ff3"} Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.066826 4867 generic.go:334] "Generic (PLEG): container finished" podID="a8d05d49-5e34-414b-b449-08464cf7d5be" containerID="6f7cfedf70229812a59a62dbe096cfa62c008c306798b0f85087d7cb1fe6fadc" exitCode=0 Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.066897 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a8d05d49-5e34-414b-b449-08464cf7d5be","Type":"ContainerDied","Data":"6f7cfedf70229812a59a62dbe096cfa62c008c306798b0f85087d7cb1fe6fadc"} Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.067752 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.068013 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.068864 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.069909 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.070494 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f" exitCode=0 Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.070513 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3" exitCode=0 Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.070522 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24" exitCode=0 Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.070529 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2" exitCode=2 Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.070571 4867 scope.go:117] "RemoveContainer" containerID="1d937baf6deef72078ddf105fde0861e1e590bcc72531e7f99f108836f4b47d2" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.071743 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c"} Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.072143 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:25 crc kubenswrapper[4867]: I1212 06:52:25.072399 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.087691 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.330592 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.331128 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.331488 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460310 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir\") pod \"a8d05d49-5e34-414b-b449-08464cf7d5be\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460386 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a8d05d49-5e34-414b-b449-08464cf7d5be" (UID: "a8d05d49-5e34-414b-b449-08464cf7d5be"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460407 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock\") pod \"a8d05d49-5e34-414b-b449-08464cf7d5be\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460444 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock" (OuterVolumeSpecName: "var-lock") pod "a8d05d49-5e34-414b-b449-08464cf7d5be" (UID: "a8d05d49-5e34-414b-b449-08464cf7d5be"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460531 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access\") pod \"a8d05d49-5e34-414b-b449-08464cf7d5be\" (UID: \"a8d05d49-5e34-414b-b449-08464cf7d5be\") " Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460920 4867 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.460940 4867 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a8d05d49-5e34-414b-b449-08464cf7d5be-var-lock\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.466445 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a8d05d49-5e34-414b-b449-08464cf7d5be" (UID: "a8d05d49-5e34-414b-b449-08464cf7d5be"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:26 crc kubenswrapper[4867]: I1212 06:52:26.562483 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8d05d49-5e34-414b-b449-08464cf7d5be-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.067712 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.069029 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.069619 4867 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.069999 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.070321 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.096534 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.097079 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817" exitCode=0 Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.097143 4867 scope.go:117] "RemoveContainer" containerID="23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.097161 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.098686 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a8d05d49-5e34-414b-b449-08464cf7d5be","Type":"ContainerDied","Data":"64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9"} Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.098734 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64916daacf82dbd19a716aa0021b22ee2aff830f1fbe372ca28608d3c0b5ffe9" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.099356 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.102535 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.103032 4867 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.103493 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.113794 4867 scope.go:117] "RemoveContainer" containerID="7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.124521 4867 scope.go:117] "RemoveContainer" containerID="2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.133921 4867 scope.go:117] "RemoveContainer" containerID="6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.147909 4867 scope.go:117] "RemoveContainer" containerID="d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.160553 4867 scope.go:117] "RemoveContainer" containerID="fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169202 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169328 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169360 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169450 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169492 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169804 4867 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169878 4867 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.169904 4867 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.178253 4867 scope.go:117] "RemoveContainer" containerID="23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.178722 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\": container with ID starting with 23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f not found: ID does not exist" containerID="23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.178755 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f"} err="failed to get container status \"23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\": rpc error: code = NotFound desc = could not find container \"23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f\": container with ID starting with 23cab05982f519e66286dc581413f73f6cb14e480b79ade9875d519692ae6e5f not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.178778 4867 scope.go:117] "RemoveContainer" containerID="7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.179300 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\": container with ID starting with 7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3 not found: ID does not exist" containerID="7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.179350 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3"} err="failed to get container status \"7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\": rpc error: code = NotFound desc = could not find container \"7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3\": container with ID starting with 7306f31b676b657205a94937637284de7bf4da96037085e1720ceff4c245bde3 not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.179379 4867 scope.go:117] "RemoveContainer" containerID="2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.179702 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\": container with ID starting with 2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24 not found: ID does not exist" containerID="2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.179753 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24"} err="failed to get container status \"2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\": rpc error: code = NotFound desc = could not find container \"2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24\": container with ID starting with 2dd08db1a022621e087570a9bac449542f84ed62a8ad08b7596df9945f7ccc24 not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.179779 4867 scope.go:117] "RemoveContainer" containerID="6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.180173 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\": container with ID starting with 6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2 not found: ID does not exist" containerID="6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.180202 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2"} err="failed to get container status \"6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\": rpc error: code = NotFound desc = could not find container \"6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2\": container with ID starting with 6c13495e072683b9b7811b9b8b6482ee387fed7226f10fd163d7926ada1a8ef2 not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.180218 4867 scope.go:117] "RemoveContainer" containerID="d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.180705 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\": container with ID starting with d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817 not found: ID does not exist" containerID="d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.180788 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817"} err="failed to get container status \"d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\": rpc error: code = NotFound desc = could not find container \"d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817\": container with ID starting with d35f190500c67e6ef6f671177877681fa0245a33ffb6f7a4a4b95d5fb0a55817 not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.180817 4867 scope.go:117] "RemoveContainer" containerID="fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.181128 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\": container with ID starting with fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8 not found: ID does not exist" containerID="fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.181170 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8"} err="failed to get container status \"fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\": rpc error: code = NotFound desc = could not find container \"fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8\": container with ID starting with fac75414c9c318b6e2684e17bd148e5d70e1a4fca5f984111e2a8e3ca3b832e8 not found: ID does not exist" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.368219 4867 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.368894 4867 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.369519 4867 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.369796 4867 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.370116 4867 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.370159 4867 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.370537 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="200ms" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.414923 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.415358 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: I1212 06:52:27.415860 4867 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.572072 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="400ms" Dec 12 06:52:27 crc kubenswrapper[4867]: E1212 06:52:27.973350 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="800ms" Dec 12 06:52:28 crc kubenswrapper[4867]: E1212 06:52:28.774897 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="1.6s" Dec 12 06:52:28 crc kubenswrapper[4867]: I1212 06:52:28.844370 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 12 06:52:29 crc kubenswrapper[4867]: E1212 06:52:29.771346 4867 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.66:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18806537eabce10b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-12 06:52:24.765088011 +0000 UTC m=+232.336469280,LastTimestamp:2025-12-12 06:52:24.765088011 +0000 UTC m=+232.336469280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 12 06:52:30 crc kubenswrapper[4867]: E1212 06:52:30.376271 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="3.2s" Dec 12 06:52:30 crc kubenswrapper[4867]: E1212 06:52:30.827604 4867 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.66:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18806537eabce10b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-12 06:52:24.765088011 +0000 UTC m=+232.336469280,LastTimestamp:2025-12-12 06:52:24.765088011 +0000 UTC m=+232.336469280,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 12 06:52:32 crc kubenswrapper[4867]: I1212 06:52:32.840154 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:32 crc kubenswrapper[4867]: I1212 06:52:32.840987 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:33 crc kubenswrapper[4867]: E1212 06:52:33.577612 4867 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.66:6443: connect: connection refused" interval="6.4s" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.023088 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" containerName="oauth-openshift" containerID="cri-o://ec95f6abf8f5cbab7c5bf3b6f77076d5cb80a96c6329f094bdaff3c0be175b4f" gracePeriod=15 Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.142173 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" containerID="ec95f6abf8f5cbab7c5bf3b6f77076d5cb80a96c6329f094bdaff3c0be175b4f" exitCode=0 Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.142220 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" event={"ID":"8c966cee-93bd-41b1-885c-e46c0f9ede1a","Type":"ContainerDied","Data":"ec95f6abf8f5cbab7c5bf3b6f77076d5cb80a96c6329f094bdaff3c0be175b4f"} Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.409693 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.410572 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.411117 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.411617 4867 status_manager.go:851] "Failed to get status for pod" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-bfk5k\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473421 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473481 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473501 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g2kq\" (UniqueName: \"kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473518 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473537 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473557 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473589 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473617 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473647 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473681 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473702 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473720 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473743 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473761 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error\") pod \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\" (UID: \"8c966cee-93bd-41b1-885c-e46c0f9ede1a\") " Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.473997 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.474408 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.474734 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.475032 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.475195 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.479285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.479456 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq" (OuterVolumeSpecName: "kube-api-access-9g2kq") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "kube-api-access-9g2kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.479638 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.484442 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.485273 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.487561 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.487770 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.488008 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.488064 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8c966cee-93bd-41b1-885c-e46c0f9ede1a" (UID: "8c966cee-93bd-41b1-885c-e46c0f9ede1a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574885 4867 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574920 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574932 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574942 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574951 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g2kq\" (UniqueName: \"kubernetes.io/projected/8c966cee-93bd-41b1-885c-e46c0f9ede1a-kube-api-access-9g2kq\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574962 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574972 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574981 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574991 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.574999 4867 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c966cee-93bd-41b1-885c-e46c0f9ede1a-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.575008 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.575017 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.575026 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:35 crc kubenswrapper[4867]: I1212 06:52:35.575034 4867 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c966cee-93bd-41b1-885c-e46c0f9ede1a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.155787 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" event={"ID":"8c966cee-93bd-41b1-885c-e46c0f9ede1a","Type":"ContainerDied","Data":"a8b9376b0024167246e1ef55c941c610ae1cb31976ddcedd703850112c9c53a6"} Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.155860 4867 scope.go:117] "RemoveContainer" containerID="ec95f6abf8f5cbab7c5bf3b6f77076d5cb80a96c6329f094bdaff3c0be175b4f" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.155933 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.156848 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.157142 4867 status_manager.go:851] "Failed to get status for pod" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-bfk5k\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.157377 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.168402 4867 status_manager.go:851] "Failed to get status for pod" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-bfk5k\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.169054 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.169493 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.837648 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.843419 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.843908 4867 status_manager.go:851] "Failed to get status for pod" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-bfk5k\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.844421 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.861052 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.861086 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:36 crc kubenswrapper[4867]: E1212 06:52:36.861500 4867 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:36 crc kubenswrapper[4867]: I1212 06:52:36.862095 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.164480 4867 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="bccfb4375be53e5e798ba35ba3d0f8844cd11e9f98af27e9e9a0b7c517394b11" exitCode=0 Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.164614 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"bccfb4375be53e5e798ba35ba3d0f8844cd11e9f98af27e9e9a0b7c517394b11"} Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.164772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"65731600c427e0b7577219a8b4d9a8d53fad3efffb33866213a14890c734a44c"} Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.165085 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.165098 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:37 crc kubenswrapper[4867]: E1212 06:52:37.165453 4867 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.165980 4867 status_manager.go:851] "Failed to get status for pod" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" pod="openshift-authentication/oauth-openshift-558db77b4-bfk5k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-bfk5k\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.166296 4867 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:37 crc kubenswrapper[4867]: I1212 06:52:37.166673 4867 status_manager.go:851] "Failed to get status for pod" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.66:6443: connect: connection refused" Dec 12 06:52:38 crc kubenswrapper[4867]: I1212 06:52:38.174144 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5ccb51dbffbc64a9ce20bcfd656d6ecea86cf27e347bf0d3d12df8cea4d0fc85"} Dec 12 06:52:38 crc kubenswrapper[4867]: I1212 06:52:38.174208 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7ff82466011db906d736420de07a82ab63211aa7bb325adb166f56f47547927d"} Dec 12 06:52:38 crc kubenswrapper[4867]: I1212 06:52:38.174238 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6f277cb937a254f3b4c3541c1ef2125cc4a57aaf3ea35be265fc0b5e6b2497e9"} Dec 12 06:52:38 crc kubenswrapper[4867]: I1212 06:52:38.174251 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b83dc967b2a612d47d4aa0c21b732bd4d8fb52bdf716696b9da219c7b0129feb"} Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.180912 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.180953 4867 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92" exitCode=1 Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.181002 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92"} Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.181509 4867 scope.go:117] "RemoveContainer" containerID="7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92" Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.185691 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5a0d927a0e120880bfb6388f787d4ca2664e76cfdbc8af7838bfaa9605d0d9c5"} Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.185883 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.185901 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:39 crc kubenswrapper[4867]: I1212 06:52:39.186037 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:40 crc kubenswrapper[4867]: I1212 06:52:40.193581 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 12 06:52:40 crc kubenswrapper[4867]: I1212 06:52:40.193886 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6cbd9d86dc21dcdb70eb2477ddea755db3f3c2c732ca54a04d4009f83e3809ef"} Dec 12 06:52:40 crc kubenswrapper[4867]: I1212 06:52:40.870914 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:52:41 crc kubenswrapper[4867]: I1212 06:52:41.862745 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:41 crc kubenswrapper[4867]: I1212 06:52:41.863098 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:41 crc kubenswrapper[4867]: I1212 06:52:41.870120 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:44 crc kubenswrapper[4867]: I1212 06:52:44.205653 4867 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:44 crc kubenswrapper[4867]: I1212 06:52:44.322570 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9fe7ac7a-d6af-46b3-8f11-109432d1188b" Dec 12 06:52:45 crc kubenswrapper[4867]: I1212 06:52:45.218441 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:45 crc kubenswrapper[4867]: I1212 06:52:45.219388 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:45 crc kubenswrapper[4867]: I1212 06:52:45.221732 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9fe7ac7a-d6af-46b3-8f11-109432d1188b" Dec 12 06:52:45 crc kubenswrapper[4867]: I1212 06:52:45.223087 4867 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://b83dc967b2a612d47d4aa0c21b732bd4d8fb52bdf716696b9da219c7b0129feb" Dec 12 06:52:45 crc kubenswrapper[4867]: I1212 06:52:45.223103 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:52:46 crc kubenswrapper[4867]: I1212 06:52:46.222963 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:46 crc kubenswrapper[4867]: I1212 06:52:46.223005 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:52:46 crc kubenswrapper[4867]: I1212 06:52:46.226193 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9fe7ac7a-d6af-46b3-8f11-109432d1188b" Dec 12 06:52:47 crc kubenswrapper[4867]: I1212 06:52:47.239040 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:52:47 crc kubenswrapper[4867]: I1212 06:52:47.239327 4867 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 12 06:52:47 crc kubenswrapper[4867]: I1212 06:52:47.239381 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 12 06:52:52 crc kubenswrapper[4867]: I1212 06:52:52.900584 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 12 06:52:53 crc kubenswrapper[4867]: I1212 06:52:53.541156 4867 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 12 06:52:55 crc kubenswrapper[4867]: I1212 06:52:55.071758 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 12 06:52:55 crc kubenswrapper[4867]: I1212 06:52:55.145812 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 12 06:52:55 crc kubenswrapper[4867]: I1212 06:52:55.452128 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 12 06:52:55 crc kubenswrapper[4867]: I1212 06:52:55.589100 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.306594 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.327028 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.369661 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.432569 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.509271 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 12 06:52:56 crc kubenswrapper[4867]: I1212 06:52:56.936674 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.182933 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.239803 4867 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.239863 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.309537 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.481657 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.527654 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.701884 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 12 06:52:57 crc kubenswrapper[4867]: I1212 06:52:57.857032 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.021680 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.348243 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.666915 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.794157 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.842654 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.933235 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.941762 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 12 06:52:58 crc kubenswrapper[4867]: I1212 06:52:58.975352 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.118247 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.409051 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.485953 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.517581 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.558192 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.563607 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.596429 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.617780 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.703510 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.726077 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.776535 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.907697 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.920033 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.931932 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.953139 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 12 06:52:59 crc kubenswrapper[4867]: I1212 06:52:59.976289 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.066540 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.138629 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.142939 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.346920 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.373555 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.526748 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.620033 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.676631 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.698499 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.745839 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 12 06:53:00 crc kubenswrapper[4867]: I1212 06:53:00.924195 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.117057 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.184033 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.244110 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.335810 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.431461 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.431662 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.473421 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.549145 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.559672 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.612474 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.615887 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.735078 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.846426 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 12 06:53:01 crc kubenswrapper[4867]: I1212 06:53:01.968830 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.131318 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.337720 4867 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.403042 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.562111 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.645290 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.680091 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.692486 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.740472 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.756582 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.777556 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.829705 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.868319 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.876991 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.890573 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.907310 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.954287 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.964872 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.978155 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 12 06:53:02 crc kubenswrapper[4867]: I1212 06:53:02.984179 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.045768 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.077834 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.109584 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.124761 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.163287 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.168320 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.195576 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.271126 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.378340 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.411644 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.434371 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.469805 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.495341 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.567727 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.601536 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.615726 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.706033 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.741274 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.841030 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.855911 4867 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 12 06:53:03 crc kubenswrapper[4867]: I1212 06:53:03.863528 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.015445 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.163469 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.234005 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.257102 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.330916 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.361665 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.383450 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.386609 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.427659 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.482815 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.533031 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.608654 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.619216 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.626528 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.650622 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.651503 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.673274 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.702390 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.751675 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.813243 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.863875 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.890931 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 12 06:53:04 crc kubenswrapper[4867]: I1212 06:53:04.968903 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.078330 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.120207 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.135647 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.264428 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.278787 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.455495 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.486623 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.519443 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.535598 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.554422 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.705838 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.724395 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.769735 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.793329 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.912953 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.916407 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 12 06:53:05 crc kubenswrapper[4867]: I1212 06:53:05.983134 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.025984 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.032371 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.148259 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.188034 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.232981 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.289855 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.354853 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.418081 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.422045 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.441366 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.509553 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.509771 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.562193 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.685923 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.688649 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.753896 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.848385 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.884642 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.915469 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 12 06:53:06 crc kubenswrapper[4867]: I1212 06:53:06.963132 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.088382 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.090839 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.146582 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.202587 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.239901 4867 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.240010 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.240083 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.240935 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"6cbd9d86dc21dcdb70eb2477ddea755db3f3c2c732ca54a04d4009f83e3809ef"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.241343 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://6cbd9d86dc21dcdb70eb2477ddea755db3f3c2c732ca54a04d4009f83e3809ef" gracePeriod=30 Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.329990 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.335418 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.616422 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.682721 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.709533 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.751920 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.830773 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.854938 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.921722 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.933600 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 12 06:53:07 crc kubenswrapper[4867]: I1212 06:53:07.965328 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.053552 4867 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.053621 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.098528 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.156336 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.194734 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.256650 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.301679 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.366381 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.419029 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.447247 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.471478 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.497054 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.573427 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.635687 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.697486 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.751391 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.878135 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.948644 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.967151 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 12 06:53:08 crc kubenswrapper[4867]: I1212 06:53:08.980465 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.020084 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.044536 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.103071 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.256138 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.313823 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.325863 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.632620 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.665035 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 12 06:53:09 crc kubenswrapper[4867]: I1212 06:53:09.734974 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.028374 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.160846 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.203267 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.207023 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.222576 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.240736 4867 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.353497 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.454971 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.791505 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.864901 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 12 06:53:10 crc kubenswrapper[4867]: I1212 06:53:10.869864 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.011122 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.117876 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.248290 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.307854 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.622123 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.750720 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.782377 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.890637 4867 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.893593 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=48.893572986 podStartE2EDuration="48.893572986s" podCreationTimestamp="2025-12-12 06:52:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:52:44.238597074 +0000 UTC m=+251.809978353" watchObservedRunningTime="2025-12-12 06:53:11.893572986 +0000 UTC m=+279.464954255" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896290 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bfk5k","openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896392 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-848ffdc94b-ksnm8","openshift-kube-apiserver/kube-apiserver-crc"] Dec 12 06:53:11 crc kubenswrapper[4867]: E1212 06:53:11.896683 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" containerName="oauth-openshift" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896704 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" containerName="oauth-openshift" Dec 12 06:53:11 crc kubenswrapper[4867]: E1212 06:53:11.896718 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" containerName="installer" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896725 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" containerName="installer" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896726 4867 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896743 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4cb29ed6-1017-4d04-a984-ba197f53ef44" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896817 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" containerName="oauth-openshift" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.896829 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d05d49-5e34-414b-b449-08464cf7d5be" containerName="installer" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.897613 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.899596 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.900603 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.900760 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.901707 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.901795 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.901832 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.902179 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.902805 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.903089 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.903447 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.903711 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.904291 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.904498 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.910179 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.912474 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.920868 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 12 06:53:11 crc kubenswrapper[4867]: I1212 06:53:11.942596 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=27.942581333 podStartE2EDuration="27.942581333s" podCreationTimestamp="2025-12-12 06:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:53:11.938152114 +0000 UTC m=+279.509533383" watchObservedRunningTime="2025-12-12 06:53:11.942581333 +0000 UTC m=+279.513962602" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.017911 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-error\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.017970 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018002 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018029 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018416 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-policies\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018577 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018631 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018690 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018870 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018919 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2qc4\" (UniqueName: \"kubernetes.io/projected/8bea03eb-56c8-4402-b4c1-4f819054c6c9-kube-api-access-b2qc4\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.018958 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-session\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.019209 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-login\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.019351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-dir\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.019490 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.028073 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.085106 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.114364 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.120389 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-session\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121009 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-login\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121197 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-dir\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121418 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121617 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-error\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121311 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-dir\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121788 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121920 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.121975 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.122023 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-policies\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124360 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124458 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124478 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-audit-policies\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124523 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124620 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.124661 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2qc4\" (UniqueName: \"kubernetes.io/projected/8bea03eb-56c8-4402-b4c1-4f819054c6c9-kube-api-access-b2qc4\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.125360 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-service-ca\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.125687 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.126091 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.129020 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-session\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.129031 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-router-certs\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.129110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.130659 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.130779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.130857 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.132652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-error\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.134828 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8bea03eb-56c8-4402-b4c1-4f819054c6c9-v4-0-config-user-template-login\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.144268 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.152657 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2qc4\" (UniqueName: \"kubernetes.io/projected/8bea03eb-56c8-4402-b4c1-4f819054c6c9-kube-api-access-b2qc4\") pod \"oauth-openshift-848ffdc94b-ksnm8\" (UID: \"8bea03eb-56c8-4402-b4c1-4f819054c6c9\") " pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.217444 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.351136 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.421115 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-848ffdc94b-ksnm8"] Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.847748 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c966cee-93bd-41b1-885c-e46c0f9ede1a" path="/var/lib/kubelet/pods/8c966cee-93bd-41b1-885c-e46c0f9ede1a/volumes" Dec 12 06:53:12 crc kubenswrapper[4867]: I1212 06:53:12.898744 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 12 06:53:13 crc kubenswrapper[4867]: I1212 06:53:13.361864 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" event={"ID":"8bea03eb-56c8-4402-b4c1-4f819054c6c9","Type":"ContainerStarted","Data":"6a11524a7908b4992fa683fe8b26453e4f2f07b1754ed022fc52acdbac16e404"} Dec 12 06:53:13 crc kubenswrapper[4867]: I1212 06:53:13.362885 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" event={"ID":"8bea03eb-56c8-4402-b4c1-4f819054c6c9","Type":"ContainerStarted","Data":"ccd87230dc44ed6de659ec964ccdc56736e8e746076e3a552fa9bb1dedae4083"} Dec 12 06:53:13 crc kubenswrapper[4867]: I1212 06:53:13.387167 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" podStartSLOduration=63.38715004 podStartE2EDuration="1m3.38715004s" podCreationTimestamp="2025-12-12 06:52:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:53:13.385284375 +0000 UTC m=+280.956665664" watchObservedRunningTime="2025-12-12 06:53:13.38715004 +0000 UTC m=+280.958531309" Dec 12 06:53:13 crc kubenswrapper[4867]: I1212 06:53:13.684758 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 12 06:53:14 crc kubenswrapper[4867]: I1212 06:53:14.369718 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:14 crc kubenswrapper[4867]: I1212 06:53:14.376277 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-848ffdc94b-ksnm8" Dec 12 06:53:17 crc kubenswrapper[4867]: I1212 06:53:17.692980 4867 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 12 06:53:17 crc kubenswrapper[4867]: I1212 06:53:17.693715 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c" gracePeriod=5 Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.277468 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.277966 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.412419 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.412522 4867 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c" exitCode=137 Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.412591 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.412614 4867 scope.go:117] "RemoveContainer" containerID="53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.426547 4867 scope.go:117] "RemoveContainer" containerID="53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c" Dec 12 06:53:23 crc kubenswrapper[4867]: E1212 06:53:23.426893 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c\": container with ID starting with 53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c not found: ID does not exist" containerID="53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.426935 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c"} err="failed to get container status \"53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c\": rpc error: code = NotFound desc = could not find container \"53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c\": container with ID starting with 53d1aa608db875e4c2b5bb0b16fbfee02f298d8e6719f31a2adb765d18b9220c not found: ID does not exist" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465415 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465495 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465533 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465582 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465606 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465612 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465679 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465724 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.465745 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.466110 4867 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.466138 4867 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.466156 4867 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.466173 4867 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.477480 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 06:53:23 crc kubenswrapper[4867]: I1212 06:53:23.566961 4867 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.843849 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.844312 4867 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.855926 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.855995 4867 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f6955828-2859-44f6-aa0c-1734f19383ef" Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.860153 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 12 06:53:24 crc kubenswrapper[4867]: I1212 06:53:24.860178 4867 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="f6955828-2859-44f6-aa0c-1734f19383ef" Dec 12 06:53:27 crc kubenswrapper[4867]: I1212 06:53:27.436124 4867 generic.go:334] "Generic (PLEG): container finished" podID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerID="eaae5e77d81d72cea32b6e76537f85ff052ddc2ea64e432f6fe4c99f878eddf2" exitCode=0 Dec 12 06:53:27 crc kubenswrapper[4867]: I1212 06:53:27.436407 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerDied","Data":"eaae5e77d81d72cea32b6e76537f85ff052ddc2ea64e432f6fe4c99f878eddf2"} Dec 12 06:53:27 crc kubenswrapper[4867]: I1212 06:53:27.437231 4867 scope.go:117] "RemoveContainer" containerID="eaae5e77d81d72cea32b6e76537f85ff052ddc2ea64e432f6fe4c99f878eddf2" Dec 12 06:53:28 crc kubenswrapper[4867]: I1212 06:53:28.443832 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerStarted","Data":"c7f0e8b1f485860f43804ed5a8ea25371c6e0e584710de49618a5cb1b459e90b"} Dec 12 06:53:28 crc kubenswrapper[4867]: I1212 06:53:28.444686 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:53:28 crc kubenswrapper[4867]: I1212 06:53:28.446626 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.064937 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.065867 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerName="controller-manager" containerID="cri-o://a8929b2dc908ead36c3325990a96c8e7e5c73b8ef8d50d2f338edb4b423b272d" gracePeriod=30 Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.069702 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.069924 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerName="route-controller-manager" containerID="cri-o://823aaef7d3679c19382e449d2b9a5a00ad525209d66d1b1310b491103bce4f6b" gracePeriod=30 Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.465993 4867 generic.go:334] "Generic (PLEG): container finished" podID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerID="823aaef7d3679c19382e449d2b9a5a00ad525209d66d1b1310b491103bce4f6b" exitCode=0 Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.466087 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" event={"ID":"37ad992d-0e36-434e-9d11-106ef955c5dd","Type":"ContainerDied","Data":"823aaef7d3679c19382e449d2b9a5a00ad525209d66d1b1310b491103bce4f6b"} Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.466384 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" event={"ID":"37ad992d-0e36-434e-9d11-106ef955c5dd","Type":"ContainerDied","Data":"2f27a33585dc01956d4da1e55a919e4a27bca56559f9a5aad5f8fd52418450f9"} Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.466405 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f27a33585dc01956d4da1e55a919e4a27bca56559f9a5aad5f8fd52418450f9" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.467821 4867 generic.go:334] "Generic (PLEG): container finished" podID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerID="a8929b2dc908ead36c3325990a96c8e7e5c73b8ef8d50d2f338edb4b423b272d" exitCode=0 Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.467863 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" event={"ID":"32c762a6-9bfc-40d6-8a78-eb4189e885bd","Type":"ContainerDied","Data":"a8929b2dc908ead36c3325990a96c8e7e5c73b8ef8d50d2f338edb4b423b272d"} Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.467886 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" event={"ID":"32c762a6-9bfc-40d6-8a78-eb4189e885bd","Type":"ContainerDied","Data":"a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe"} Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.467898 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72de2863f2ee644796be08ad31644ca942cb1bfecaf43a5e4e27ec8b98749fe" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.485080 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.488961 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.678901 4867 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685497 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles\") pod \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685553 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert\") pod \"37ad992d-0e36-434e-9d11-106ef955c5dd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685577 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sd9l\" (UniqueName: \"kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l\") pod \"37ad992d-0e36-434e-9d11-106ef955c5dd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685595 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config\") pod \"37ad992d-0e36-434e-9d11-106ef955c5dd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685615 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca\") pod \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685637 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config\") pod \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685674 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28mcc\" (UniqueName: \"kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc\") pod \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert\") pod \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\" (UID: \"32c762a6-9bfc-40d6-8a78-eb4189e885bd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.685729 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca\") pod \"37ad992d-0e36-434e-9d11-106ef955c5dd\" (UID: \"37ad992d-0e36-434e-9d11-106ef955c5dd\") " Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.686407 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca" (OuterVolumeSpecName: "client-ca") pod "37ad992d-0e36-434e-9d11-106ef955c5dd" (UID: "37ad992d-0e36-434e-9d11-106ef955c5dd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.686474 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config" (OuterVolumeSpecName: "config") pod "37ad992d-0e36-434e-9d11-106ef955c5dd" (UID: "37ad992d-0e36-434e-9d11-106ef955c5dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.686749 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca" (OuterVolumeSpecName: "client-ca") pod "32c762a6-9bfc-40d6-8a78-eb4189e885bd" (UID: "32c762a6-9bfc-40d6-8a78-eb4189e885bd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.686824 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "32c762a6-9bfc-40d6-8a78-eb4189e885bd" (UID: "32c762a6-9bfc-40d6-8a78-eb4189e885bd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.687113 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config" (OuterVolumeSpecName: "config") pod "32c762a6-9bfc-40d6-8a78-eb4189e885bd" (UID: "32c762a6-9bfc-40d6-8a78-eb4189e885bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.692255 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l" (OuterVolumeSpecName: "kube-api-access-4sd9l") pod "37ad992d-0e36-434e-9d11-106ef955c5dd" (UID: "37ad992d-0e36-434e-9d11-106ef955c5dd"). InnerVolumeSpecName "kube-api-access-4sd9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.693439 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc" (OuterVolumeSpecName: "kube-api-access-28mcc") pod "32c762a6-9bfc-40d6-8a78-eb4189e885bd" (UID: "32c762a6-9bfc-40d6-8a78-eb4189e885bd"). InnerVolumeSpecName "kube-api-access-28mcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.695944 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "37ad992d-0e36-434e-9d11-106ef955c5dd" (UID: "37ad992d-0e36-434e-9d11-106ef955c5dd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.698062 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "32c762a6-9bfc-40d6-8a78-eb4189e885bd" (UID: "32c762a6-9bfc-40d6-8a78-eb4189e885bd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787416 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28mcc\" (UniqueName: \"kubernetes.io/projected/32c762a6-9bfc-40d6-8a78-eb4189e885bd-kube-api-access-28mcc\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787450 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c762a6-9bfc-40d6-8a78-eb4189e885bd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787461 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787469 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787476 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37ad992d-0e36-434e-9d11-106ef955c5dd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787484 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sd9l\" (UniqueName: \"kubernetes.io/projected/37ad992d-0e36-434e-9d11-106ef955c5dd-kube-api-access-4sd9l\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787494 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37ad992d-0e36-434e-9d11-106ef955c5dd-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787507 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:32 crc kubenswrapper[4867]: I1212 06:53:32.787520 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c762a6-9bfc-40d6-8a78-eb4189e885bd-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.472298 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-45697" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.472298 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.488927 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.494531 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-g9mhd"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.506511 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.510545 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-45697"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.882729 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:53:33 crc kubenswrapper[4867]: E1212 06:53:33.883050 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883076 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 12 06:53:33 crc kubenswrapper[4867]: E1212 06:53:33.883100 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerName="controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883109 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerName="controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: E1212 06:53:33.883127 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerName="route-controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883137 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerName="route-controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883305 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883326 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" containerName="controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883352 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" containerName="route-controller-manager" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.883876 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.885817 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.886546 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.886734 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.888528 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.888909 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.889074 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.889640 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.889811 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.890207 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.890318 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.890375 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.890398 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.890426 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.898879 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.903829 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.904282 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.940164 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.998902 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.998955 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.998994 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999029 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blzst\" (UniqueName: \"kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999066 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999093 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999139 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999161 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:33 crc kubenswrapper[4867]: I1212 06:53:33.999183 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl7fl\" (UniqueName: \"kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100292 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100350 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100381 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100402 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blzst\" (UniqueName: \"kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100426 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100451 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100485 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100501 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.100516 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl7fl\" (UniqueName: \"kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.102598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.102936 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.103150 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.103856 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.104119 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.107559 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.107810 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.116128 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl7fl\" (UniqueName: \"kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl\") pod \"controller-manager-58bf4dd977-7rnkr\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.118920 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blzst\" (UniqueName: \"kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst\") pod \"route-controller-manager-7cbb74dfd9-dp4l7\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.210887 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.219734 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.624308 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.672319 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:53:34 crc kubenswrapper[4867]: W1212 06:53:34.676175 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c44bf3_87fc_41fe_b953_291e1b8f8b6a.slice/crio-c58a1b5120ce3e850c187e603c8e1e32b6488914403ef9b705f1fbc1b059943b WatchSource:0}: Error finding container c58a1b5120ce3e850c187e603c8e1e32b6488914403ef9b705f1fbc1b059943b: Status 404 returned error can't find the container with id c58a1b5120ce3e850c187e603c8e1e32b6488914403ef9b705f1fbc1b059943b Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.849598 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c762a6-9bfc-40d6-8a78-eb4189e885bd" path="/var/lib/kubelet/pods/32c762a6-9bfc-40d6-8a78-eb4189e885bd/volumes" Dec 12 06:53:34 crc kubenswrapper[4867]: I1212 06:53:34.850362 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ad992d-0e36-434e-9d11-106ef955c5dd" path="/var/lib/kubelet/pods/37ad992d-0e36-434e-9d11-106ef955c5dd/volumes" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.481669 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" event={"ID":"07c44bf3-87fc-41fe-b953-291e1b8f8b6a","Type":"ContainerStarted","Data":"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6"} Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.481997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" event={"ID":"07c44bf3-87fc-41fe-b953-291e1b8f8b6a","Type":"ContainerStarted","Data":"c58a1b5120ce3e850c187e603c8e1e32b6488914403ef9b705f1fbc1b059943b"} Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.482273 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.484032 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" event={"ID":"090c2564-d726-4214-b0c3-578018da0b8e","Type":"ContainerStarted","Data":"f51504d9217a0046b2c38c3b8a9a7efe9371cd3925b3b636c11c1d4c7bf2df00"} Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.484058 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" event={"ID":"090c2564-d726-4214-b0c3-578018da0b8e","Type":"ContainerStarted","Data":"494b670fb6fb6b21bc15a7bcd2c8326f61deb20ee81b32cc0ce331c5d0af0fd8"} Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.484312 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.488730 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.489342 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.500136 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" podStartSLOduration=3.500116932 podStartE2EDuration="3.500116932s" podCreationTimestamp="2025-12-12 06:53:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:53:35.499315812 +0000 UTC m=+303.070697081" watchObservedRunningTime="2025-12-12 06:53:35.500116932 +0000 UTC m=+303.071498201" Dec 12 06:53:35 crc kubenswrapper[4867]: I1212 06:53:35.536820 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" podStartSLOduration=3.536803898 podStartE2EDuration="3.536803898s" podCreationTimestamp="2025-12-12 06:53:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:53:35.53444964 +0000 UTC m=+303.105830919" watchObservedRunningTime="2025-12-12 06:53:35.536803898 +0000 UTC m=+303.108185167" Dec 12 06:53:37 crc kubenswrapper[4867]: I1212 06:53:37.494466 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 12 06:53:37 crc kubenswrapper[4867]: I1212 06:53:37.496476 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 12 06:53:37 crc kubenswrapper[4867]: I1212 06:53:37.496516 4867 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6cbd9d86dc21dcdb70eb2477ddea755db3f3c2c732ca54a04d4009f83e3809ef" exitCode=137 Dec 12 06:53:37 crc kubenswrapper[4867]: I1212 06:53:37.496550 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6cbd9d86dc21dcdb70eb2477ddea755db3f3c2c732ca54a04d4009f83e3809ef"} Dec 12 06:53:37 crc kubenswrapper[4867]: I1212 06:53:37.496591 4867 scope.go:117] "RemoveContainer" containerID="7b9e47867f46d30efae5f2aa47a92172608a2ddb5b757e2c11cad2a8a80a5c92" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.135822 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.136089 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g678h" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="registry-server" containerID="cri-o://443f97c96d0ae47ac5bc0e8273d01be0309b08be5c029d10795a95a26bf06d0b" gracePeriod=30 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.139982 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.140353 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fmxsr" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="registry-server" containerID="cri-o://5af7887ed4718c2c1df5b64c9bca2c9d5bb5a905f315e01c60b8dca0437dd987" gracePeriod=30 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.154494 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.154707 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" containerID="cri-o://c7f0e8b1f485860f43804ed5a8ea25371c6e0e584710de49618a5cb1b459e90b" gracePeriod=30 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.165495 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.165771 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jpvvh" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="registry-server" containerID="cri-o://68d3194db8704b335d0e73715c75fd1171f0020a2e181511f28802454a53819f" gracePeriod=30 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.175090 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.175331 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rr6lx" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="registry-server" containerID="cri-o://a18d4983b7ed6705e7087748f849efa32e6913d48ec963fc9ce35ef67e7bdb2a" gracePeriod=30 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.503440 4867 generic.go:334] "Generic (PLEG): container finished" podID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerID="c7f0e8b1f485860f43804ed5a8ea25371c6e0e584710de49618a5cb1b459e90b" exitCode=0 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.503579 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerDied","Data":"c7f0e8b1f485860f43804ed5a8ea25371c6e0e584710de49618a5cb1b459e90b"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.503862 4867 scope.go:117] "RemoveContainer" containerID="eaae5e77d81d72cea32b6e76537f85ff052ddc2ea64e432f6fe4c99f878eddf2" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.534075 4867 generic.go:334] "Generic (PLEG): container finished" podID="f2cda128-4d99-448e-9883-b53bae064a1b" containerID="a18d4983b7ed6705e7087748f849efa32e6913d48ec963fc9ce35ef67e7bdb2a" exitCode=0 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.534133 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerDied","Data":"a18d4983b7ed6705e7087748f849efa32e6913d48ec963fc9ce35ef67e7bdb2a"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.535712 4867 generic.go:334] "Generic (PLEG): container finished" podID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerID="68d3194db8704b335d0e73715c75fd1171f0020a2e181511f28802454a53819f" exitCode=0 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.535756 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerDied","Data":"68d3194db8704b335d0e73715c75fd1171f0020a2e181511f28802454a53819f"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.537674 4867 generic.go:334] "Generic (PLEG): container finished" podID="739851dc-9291-46a6-bea7-0c89156eacd1" containerID="5af7887ed4718c2c1df5b64c9bca2c9d5bb5a905f315e01c60b8dca0437dd987" exitCode=0 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.537719 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerDied","Data":"5af7887ed4718c2c1df5b64c9bca2c9d5bb5a905f315e01c60b8dca0437dd987"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.539993 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.541090 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"96410dcb4b93ec3e300cd00de651456b70b2043f8920283596b9fd265f5ff8b7"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.559590 4867 generic.go:334] "Generic (PLEG): container finished" podID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerID="443f97c96d0ae47ac5bc0e8273d01be0309b08be5c029d10795a95a26bf06d0b" exitCode=0 Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.559634 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerDied","Data":"443f97c96d0ae47ac5bc0e8273d01be0309b08be5c029d10795a95a26bf06d0b"} Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.613999 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.625721 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.645112 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.647324 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.652636 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669468 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxs6q\" (UniqueName: \"kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q\") pod \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669514 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities\") pod \"739851dc-9291-46a6-bea7-0c89156eacd1\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669599 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zg4m\" (UniqueName: \"kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m\") pod \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669633 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities\") pod \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669651 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content\") pod \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669673 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content\") pod \"739851dc-9291-46a6-bea7-0c89156eacd1\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669702 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities\") pod \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669719 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54nv8\" (UniqueName: \"kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8\") pod \"739851dc-9291-46a6-bea7-0c89156eacd1\" (UID: \"739851dc-9291-46a6-bea7-0c89156eacd1\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669764 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wf6z\" (UniqueName: \"kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z\") pod \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\" (UID: \"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.669782 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content\") pod \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\" (UID: \"b3613595-6ad9-482c-a9fd-1a3e2c5cc430\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.671384 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca\") pod \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.671415 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content\") pod \"f2cda128-4d99-448e-9883-b53bae064a1b\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.671433 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics\") pod \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\" (UID: \"49c2edc7-6a68-47ed-bd2f-581ba68aadc7\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.671450 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities\") pod \"f2cda128-4d99-448e-9883-b53bae064a1b\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.671466 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw2kz\" (UniqueName: \"kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz\") pod \"f2cda128-4d99-448e-9883-b53bae064a1b\" (UID: \"f2cda128-4d99-448e-9883-b53bae064a1b\") " Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.672107 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities" (OuterVolumeSpecName: "utilities") pod "b3613595-6ad9-482c-a9fd-1a3e2c5cc430" (UID: "b3613595-6ad9-482c-a9fd-1a3e2c5cc430"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.672959 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "49c2edc7-6a68-47ed-bd2f-581ba68aadc7" (UID: "49c2edc7-6a68-47ed-bd2f-581ba68aadc7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.673078 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities" (OuterVolumeSpecName: "utilities") pod "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" (UID: "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.675211 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities" (OuterVolumeSpecName: "utilities") pod "f2cda128-4d99-448e-9883-b53bae064a1b" (UID: "f2cda128-4d99-448e-9883-b53bae064a1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.680056 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities" (OuterVolumeSpecName: "utilities") pod "739851dc-9291-46a6-bea7-0c89156eacd1" (UID: "739851dc-9291-46a6-bea7-0c89156eacd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.685711 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z" (OuterVolumeSpecName: "kube-api-access-7wf6z") pod "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" (UID: "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa"). InnerVolumeSpecName "kube-api-access-7wf6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.685989 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz" (OuterVolumeSpecName: "kube-api-access-gw2kz") pod "f2cda128-4d99-448e-9883-b53bae064a1b" (UID: "f2cda128-4d99-448e-9883-b53bae064a1b"). InnerVolumeSpecName "kube-api-access-gw2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.686082 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8" (OuterVolumeSpecName: "kube-api-access-54nv8") pod "739851dc-9291-46a6-bea7-0c89156eacd1" (UID: "739851dc-9291-46a6-bea7-0c89156eacd1"). InnerVolumeSpecName "kube-api-access-54nv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.686730 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q" (OuterVolumeSpecName: "kube-api-access-xxs6q") pod "49c2edc7-6a68-47ed-bd2f-581ba68aadc7" (UID: "49c2edc7-6a68-47ed-bd2f-581ba68aadc7"). InnerVolumeSpecName "kube-api-access-xxs6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.698661 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m" (OuterVolumeSpecName: "kube-api-access-8zg4m") pod "b3613595-6ad9-482c-a9fd-1a3e2c5cc430" (UID: "b3613595-6ad9-482c-a9fd-1a3e2c5cc430"). InnerVolumeSpecName "kube-api-access-8zg4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.705861 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "49c2edc7-6a68-47ed-bd2f-581ba68aadc7" (UID: "49c2edc7-6a68-47ed-bd2f-581ba68aadc7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.710802 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" (UID: "8e0f2043-4b40-47c5-96ec-0e13dafcf1fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.756326 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3613595-6ad9-482c-a9fd-1a3e2c5cc430" (UID: "b3613595-6ad9-482c-a9fd-1a3e2c5cc430"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.759693 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "739851dc-9291-46a6-bea7-0c89156eacd1" (UID: "739851dc-9291-46a6-bea7-0c89156eacd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772409 4867 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772439 4867 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772451 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772461 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw2kz\" (UniqueName: \"kubernetes.io/projected/f2cda128-4d99-448e-9883-b53bae064a1b-kube-api-access-gw2kz\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772469 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxs6q\" (UniqueName: \"kubernetes.io/projected/49c2edc7-6a68-47ed-bd2f-581ba68aadc7-kube-api-access-xxs6q\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772478 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772486 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zg4m\" (UniqueName: \"kubernetes.io/projected/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-kube-api-access-8zg4m\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772516 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772525 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772533 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/739851dc-9291-46a6-bea7-0c89156eacd1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772541 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772549 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54nv8\" (UniqueName: \"kubernetes.io/projected/739851dc-9291-46a6-bea7-0c89156eacd1-kube-api-access-54nv8\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772559 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wf6z\" (UniqueName: \"kubernetes.io/projected/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa-kube-api-access-7wf6z\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.772567 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3613595-6ad9-482c-a9fd-1a3e2c5cc430-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.818285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2cda128-4d99-448e-9883-b53bae064a1b" (UID: "f2cda128-4d99-448e-9883-b53bae064a1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:53:38 crc kubenswrapper[4867]: I1212 06:53:38.873198 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2cda128-4d99-448e-9883-b53bae064a1b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.566937 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g678h" event={"ID":"b3613595-6ad9-482c-a9fd-1a3e2c5cc430","Type":"ContainerDied","Data":"da6d9f7f41d176f8e7f039141099bab528e2721232893ca0c7568f26d979409a"} Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.567000 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g678h" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.567013 4867 scope.go:117] "RemoveContainer" containerID="443f97c96d0ae47ac5bc0e8273d01be0309b08be5c029d10795a95a26bf06d0b" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.568338 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" event={"ID":"49c2edc7-6a68-47ed-bd2f-581ba68aadc7","Type":"ContainerDied","Data":"d2ec66ee7b435fd6bc7e6a6cece258be1b306796892c12aa31d1cc0b1cfb1c33"} Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.568369 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlb8v" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.571824 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rr6lx" event={"ID":"f2cda128-4d99-448e-9883-b53bae064a1b","Type":"ContainerDied","Data":"3fec0ccd5b6350d8019904e832f071040a55f8556e42e18dd6d856188df85b9b"} Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.572062 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rr6lx" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.573929 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpvvh" event={"ID":"8e0f2043-4b40-47c5-96ec-0e13dafcf1fa","Type":"ContainerDied","Data":"d2a645b87a51fe1b9f169536da80d4716be8c486a9cd62cb30fa1806c98f0a95"} Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.573961 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpvvh" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.575903 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fmxsr" event={"ID":"739851dc-9291-46a6-bea7-0c89156eacd1","Type":"ContainerDied","Data":"fd15f3f514a883c154fda06cee0a7ad630036a769c110b1956da8b65c6e3ff54"} Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.575952 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fmxsr" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.593353 4867 scope.go:117] "RemoveContainer" containerID="fe3c2d97e73803e5d58d8b4befcbb6fb3396771206c5d7cc129d2c5b316ec03b" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.599953 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.608210 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g678h"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.612618 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.612978 4867 scope.go:117] "RemoveContainer" containerID="62f106f7bd3658a364573de0252e62305e77ac0b70eefe1cf3660a2d0d4456bf" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.616803 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlb8v"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.626310 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.636477 4867 scope.go:117] "RemoveContainer" containerID="c7f0e8b1f485860f43804ed5a8ea25371c6e0e584710de49618a5cb1b459e90b" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.649452 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpvvh"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.658059 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.659830 4867 scope.go:117] "RemoveContainer" containerID="a18d4983b7ed6705e7087748f849efa32e6913d48ec963fc9ce35ef67e7bdb2a" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.662049 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fmxsr"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.664770 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.667589 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rr6lx"] Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.674567 4867 scope.go:117] "RemoveContainer" containerID="1d7271931eb38873a9afd1afbbc7c313e057214c49c55d78d4221db8dd089e56" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.693119 4867 scope.go:117] "RemoveContainer" containerID="7ad07b38d4e5efd75cf22e8691d00bcae6f2da59b529c300017fcfcf5581653b" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.707159 4867 scope.go:117] "RemoveContainer" containerID="68d3194db8704b335d0e73715c75fd1171f0020a2e181511f28802454a53819f" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.726487 4867 scope.go:117] "RemoveContainer" containerID="d91e09ed0e36cc8d98cc648413a0d2c280df8cab6684c6b0c750f0fa3d2db814" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.740606 4867 scope.go:117] "RemoveContainer" containerID="43f6b5f3ae4b7147223839366c8fcfbca7f942d4d35580b6acccb8a2e78aad76" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.752112 4867 scope.go:117] "RemoveContainer" containerID="5af7887ed4718c2c1df5b64c9bca2c9d5bb5a905f315e01c60b8dca0437dd987" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.765035 4867 scope.go:117] "RemoveContainer" containerID="93ea830534342a60c996bc5475ee6f3e9279ce1bc534f183cbe6ae35eba36a06" Dec 12 06:53:39 crc kubenswrapper[4867]: I1212 06:53:39.780859 4867 scope.go:117] "RemoveContainer" containerID="555548831c6736123f563b9cd2011d3f76f74497e6034731108fdf7f45e940de" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.850803 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" path="/var/lib/kubelet/pods/49c2edc7-6a68-47ed-bd2f-581ba68aadc7/volumes" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.853185 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" path="/var/lib/kubelet/pods/739851dc-9291-46a6-bea7-0c89156eacd1/volumes" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.854755 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" path="/var/lib/kubelet/pods/8e0f2043-4b40-47c5-96ec-0e13dafcf1fa/volumes" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.857314 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" path="/var/lib/kubelet/pods/b3613595-6ad9-482c-a9fd-1a3e2c5cc430/volumes" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.858756 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" path="/var/lib/kubelet/pods/f2cda128-4d99-448e-9883-b53bae064a1b/volumes" Dec 12 06:53:40 crc kubenswrapper[4867]: I1212 06:53:40.871172 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.483531 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.483982 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484056 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484121 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484185 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484270 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484346 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484404 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484458 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484615 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484691 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484766 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484828 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.484884 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.484945 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.485003 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.485060 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.485163 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.485260 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.485332 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.485410 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="extract-utilities" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.485472 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.486559 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="extract-content" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.486648 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.486723 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.486821 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.486879 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.487034 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.487783 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="739851dc-9291-46a6-bea7-0c89156eacd1" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.487873 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2cda128-4d99-448e-9883-b53bae064a1b" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.487952 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0f2043-4b40-47c5-96ec-0e13dafcf1fa" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.488047 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3613595-6ad9-482c-a9fd-1a3e2c5cc430" containerName="registry-server" Dec 12 06:53:41 crc kubenswrapper[4867]: E1212 06:53:41.488217 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.488304 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.488459 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c2edc7-6a68-47ed-bd2f-581ba68aadc7" containerName="marketplace-operator" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.542349 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.545563 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.547774 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.548023 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.548781 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.601830 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.601883 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xqvp\" (UniqueName: \"kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.602084 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.703040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.703097 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.703130 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xqvp\" (UniqueName: \"kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.703975 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.703976 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.722426 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xqvp\" (UniqueName: \"kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp\") pod \"community-operators-hszdl\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:41 crc kubenswrapper[4867]: I1212 06:53:41.857391 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.265543 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 06:53:42 crc kubenswrapper[4867]: W1212 06:53:42.277527 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac8f3914_dcac_4c7d_8fa5_6c7648197f43.slice/crio-1275973da752f01890689a8dfac657ecaf3ca55a10ad60265fc6815a8a8cd651 WatchSource:0}: Error finding container 1275973da752f01890689a8dfac657ecaf3ca55a10ad60265fc6815a8a8cd651: Status 404 returned error can't find the container with id 1275973da752f01890689a8dfac657ecaf3ca55a10ad60265fc6815a8a8cd651 Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.477649 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9dng"] Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.479086 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.481461 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.485459 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9dng"] Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.511338 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-catalog-content\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.511401 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzqb5\" (UniqueName: \"kubernetes.io/projected/097fc828-65da-4d60-9aac-3ff18d583b12-kube-api-access-fzqb5\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.511426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-utilities\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.612205 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-catalog-content\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.612346 4867 generic.go:334] "Generic (PLEG): container finished" podID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerID="4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049" exitCode=0 Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.612384 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerDied","Data":"4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049"} Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.612608 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerStarted","Data":"1275973da752f01890689a8dfac657ecaf3ca55a10ad60265fc6815a8a8cd651"} Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.612656 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-catalog-content\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.613047 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzqb5\" (UniqueName: \"kubernetes.io/projected/097fc828-65da-4d60-9aac-3ff18d583b12-kube-api-access-fzqb5\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.613186 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-utilities\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.613523 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/097fc828-65da-4d60-9aac-3ff18d583b12-utilities\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.633661 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzqb5\" (UniqueName: \"kubernetes.io/projected/097fc828-65da-4d60-9aac-3ff18d583b12-kube-api-access-fzqb5\") pod \"certified-operators-p9dng\" (UID: \"097fc828-65da-4d60-9aac-3ff18d583b12\") " pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:42 crc kubenswrapper[4867]: I1212 06:53:42.831131 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.242118 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9dng"] Dec 12 06:53:43 crc kubenswrapper[4867]: W1212 06:53:43.247695 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod097fc828_65da_4d60_9aac_3ff18d583b12.slice/crio-fa7c4479d4dac9c695e6cfb1af50eaa2afbe65ce1c5ca2a71302fc4330281cf6 WatchSource:0}: Error finding container fa7c4479d4dac9c695e6cfb1af50eaa2afbe65ce1c5ca2a71302fc4330281cf6: Status 404 returned error can't find the container with id fa7c4479d4dac9c695e6cfb1af50eaa2afbe65ce1c5ca2a71302fc4330281cf6 Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.619496 4867 generic.go:334] "Generic (PLEG): container finished" podID="097fc828-65da-4d60-9aac-3ff18d583b12" containerID="eb457c141c0e61dd438fe15260bf39fec5bf8ed7e0785f8d4d1105dd319ff682" exitCode=0 Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.619841 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9dng" event={"ID":"097fc828-65da-4d60-9aac-3ff18d583b12","Type":"ContainerDied","Data":"eb457c141c0e61dd438fe15260bf39fec5bf8ed7e0785f8d4d1105dd319ff682"} Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.620130 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9dng" event={"ID":"097fc828-65da-4d60-9aac-3ff18d583b12","Type":"ContainerStarted","Data":"fa7c4479d4dac9c695e6cfb1af50eaa2afbe65ce1c5ca2a71302fc4330281cf6"} Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.624611 4867 generic.go:334] "Generic (PLEG): container finished" podID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerID="135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125" exitCode=0 Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.624681 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerDied","Data":"135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125"} Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.878612 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqrr"] Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.879877 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.881742 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.893075 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqrr"] Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.931592 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-catalog-content\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.931635 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhgrl\" (UniqueName: \"kubernetes.io/projected/b08a5662-0455-4b9e-84db-f471b8a74d33-kube-api-access-mhgrl\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:43 crc kubenswrapper[4867]: I1212 06:53:43.931689 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-utilities\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.033293 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-utilities\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.033420 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-catalog-content\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.033455 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhgrl\" (UniqueName: \"kubernetes.io/projected/b08a5662-0455-4b9e-84db-f471b8a74d33-kube-api-access-mhgrl\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.033856 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-utilities\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.034025 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b08a5662-0455-4b9e-84db-f471b8a74d33-catalog-content\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.053817 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhgrl\" (UniqueName: \"kubernetes.io/projected/b08a5662-0455-4b9e-84db-f471b8a74d33-kube-api-access-mhgrl\") pod \"redhat-marketplace-qqqrr\" (UID: \"b08a5662-0455-4b9e-84db-f471b8a74d33\") " pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.222477 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.638208 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqrr"] Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.650903 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerStarted","Data":"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63"} Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.656216 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9dng" event={"ID":"097fc828-65da-4d60-9aac-3ff18d583b12","Type":"ContainerStarted","Data":"fc357266e23a6ad9baa01e924a2fef05aba4a3aa61a34211707adf0625441b9a"} Dec 12 06:53:44 crc kubenswrapper[4867]: I1212 06:53:44.683858 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hszdl" podStartSLOduration=2.156002132 podStartE2EDuration="3.683837204s" podCreationTimestamp="2025-12-12 06:53:41 +0000 UTC" firstStartedPulling="2025-12-12 06:53:42.613648765 +0000 UTC m=+310.185030034" lastFinishedPulling="2025-12-12 06:53:44.141483837 +0000 UTC m=+311.712865106" observedRunningTime="2025-12-12 06:53:44.676393211 +0000 UTC m=+312.247774490" watchObservedRunningTime="2025-12-12 06:53:44.683837204 +0000 UTC m=+312.255218483" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.085215 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2hql5"] Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.086721 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.088202 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.099085 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2hql5"] Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.144570 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-catalog-content\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.144656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-utilities\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.144698 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d698t\" (UniqueName: \"kubernetes.io/projected/196dd20e-43b3-4f25-a61c-a94d958bb555-kube-api-access-d698t\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.245560 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-catalog-content\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.245628 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-utilities\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.245663 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d698t\" (UniqueName: \"kubernetes.io/projected/196dd20e-43b3-4f25-a61c-a94d958bb555-kube-api-access-d698t\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.246067 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-catalog-content\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.246119 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196dd20e-43b3-4f25-a61c-a94d958bb555-utilities\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.277145 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d698t\" (UniqueName: \"kubernetes.io/projected/196dd20e-43b3-4f25-a61c-a94d958bb555-kube-api-access-d698t\") pod \"redhat-operators-2hql5\" (UID: \"196dd20e-43b3-4f25-a61c-a94d958bb555\") " pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.404197 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.663143 4867 generic.go:334] "Generic (PLEG): container finished" podID="b08a5662-0455-4b9e-84db-f471b8a74d33" containerID="023e02217462afacb515055371c519b69d0c7ac9310169fed0f2f8a7cf719fa8" exitCode=0 Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.663388 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqrr" event={"ID":"b08a5662-0455-4b9e-84db-f471b8a74d33","Type":"ContainerDied","Data":"023e02217462afacb515055371c519b69d0c7ac9310169fed0f2f8a7cf719fa8"} Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.663417 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqrr" event={"ID":"b08a5662-0455-4b9e-84db-f471b8a74d33","Type":"ContainerStarted","Data":"bab33c3b87caf248ed3395f7afe7303e3ca26065d332576f98aad2844729082f"} Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.666801 4867 generic.go:334] "Generic (PLEG): container finished" podID="097fc828-65da-4d60-9aac-3ff18d583b12" containerID="fc357266e23a6ad9baa01e924a2fef05aba4a3aa61a34211707adf0625441b9a" exitCode=0 Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.667191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9dng" event={"ID":"097fc828-65da-4d60-9aac-3ff18d583b12","Type":"ContainerDied","Data":"fc357266e23a6ad9baa01e924a2fef05aba4a3aa61a34211707adf0625441b9a"} Dec 12 06:53:45 crc kubenswrapper[4867]: I1212 06:53:45.802040 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2hql5"] Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.673128 4867 generic.go:334] "Generic (PLEG): container finished" podID="196dd20e-43b3-4f25-a61c-a94d958bb555" containerID="cfa0808c3964d09812073c48f3408843dc4f5ced1bad62fe6bfbb53a8c5d030d" exitCode=0 Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.673219 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2hql5" event={"ID":"196dd20e-43b3-4f25-a61c-a94d958bb555","Type":"ContainerDied","Data":"cfa0808c3964d09812073c48f3408843dc4f5ced1bad62fe6bfbb53a8c5d030d"} Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.673490 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2hql5" event={"ID":"196dd20e-43b3-4f25-a61c-a94d958bb555","Type":"ContainerStarted","Data":"9a07aa3e8dd1d97d9b96c1a9361f2ab092a64e860679a288a7ab9c95e8701d09"} Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.675277 4867 generic.go:334] "Generic (PLEG): container finished" podID="b08a5662-0455-4b9e-84db-f471b8a74d33" containerID="88395a25c615cec6fc727b5eec5244f4b9c7615b51a15306cff8b0c201818c6d" exitCode=0 Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.675345 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqrr" event={"ID":"b08a5662-0455-4b9e-84db-f471b8a74d33","Type":"ContainerDied","Data":"88395a25c615cec6fc727b5eec5244f4b9c7615b51a15306cff8b0c201818c6d"} Dec 12 06:53:46 crc kubenswrapper[4867]: I1212 06:53:46.677712 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9dng" event={"ID":"097fc828-65da-4d60-9aac-3ff18d583b12","Type":"ContainerStarted","Data":"24782fb5559da8e3a6efd831b98f68b7ff33ca0022986872b0d0f1f34cc88035"} Dec 12 06:53:47 crc kubenswrapper[4867]: I1212 06:53:47.239197 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:53:47 crc kubenswrapper[4867]: I1212 06:53:47.242888 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:53:47 crc kubenswrapper[4867]: I1212 06:53:47.258847 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9dng" podStartSLOduration=2.809650298 podStartE2EDuration="5.25882989s" podCreationTimestamp="2025-12-12 06:53:42 +0000 UTC" firstStartedPulling="2025-12-12 06:53:43.622458932 +0000 UTC m=+311.193840201" lastFinishedPulling="2025-12-12 06:53:46.071638524 +0000 UTC m=+313.643019793" observedRunningTime="2025-12-12 06:53:46.718153605 +0000 UTC m=+314.289534874" watchObservedRunningTime="2025-12-12 06:53:47.25882989 +0000 UTC m=+314.830211159" Dec 12 06:53:47 crc kubenswrapper[4867]: I1212 06:53:47.686683 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 12 06:53:48 crc kubenswrapper[4867]: I1212 06:53:48.689884 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2hql5" event={"ID":"196dd20e-43b3-4f25-a61c-a94d958bb555","Type":"ContainerStarted","Data":"1808db55489fc4a24bae57a3a5aa003254b6c43f707617eede247ecfa67d5bef"} Dec 12 06:53:48 crc kubenswrapper[4867]: I1212 06:53:48.694215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqrr" event={"ID":"b08a5662-0455-4b9e-84db-f471b8a74d33","Type":"ContainerStarted","Data":"96c8ca0d70f405bf2a873b49f0e90733defb47a157c5593727e7d81e96dbddb8"} Dec 12 06:53:48 crc kubenswrapper[4867]: I1212 06:53:48.724248 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qqqrr" podStartSLOduration=4.334860178 podStartE2EDuration="5.724219536s" podCreationTimestamp="2025-12-12 06:53:43 +0000 UTC" firstStartedPulling="2025-12-12 06:53:45.665342633 +0000 UTC m=+313.236723902" lastFinishedPulling="2025-12-12 06:53:47.054701991 +0000 UTC m=+314.626083260" observedRunningTime="2025-12-12 06:53:48.722890194 +0000 UTC m=+316.294271463" watchObservedRunningTime="2025-12-12 06:53:48.724219536 +0000 UTC m=+316.295600805" Dec 12 06:53:49 crc kubenswrapper[4867]: I1212 06:53:49.704468 4867 generic.go:334] "Generic (PLEG): container finished" podID="196dd20e-43b3-4f25-a61c-a94d958bb555" containerID="1808db55489fc4a24bae57a3a5aa003254b6c43f707617eede247ecfa67d5bef" exitCode=0 Dec 12 06:53:49 crc kubenswrapper[4867]: I1212 06:53:49.704925 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2hql5" event={"ID":"196dd20e-43b3-4f25-a61c-a94d958bb555","Type":"ContainerDied","Data":"1808db55489fc4a24bae57a3a5aa003254b6c43f707617eede247ecfa67d5bef"} Dec 12 06:53:51 crc kubenswrapper[4867]: I1212 06:53:51.719205 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2hql5" event={"ID":"196dd20e-43b3-4f25-a61c-a94d958bb555","Type":"ContainerStarted","Data":"f3d05cae9e111676dedbd7b3417c79f3b6b93ce9527f44a896f0764c72580804"} Dec 12 06:53:51 crc kubenswrapper[4867]: I1212 06:53:51.737093 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2hql5" podStartSLOduration=2.688505862 podStartE2EDuration="6.737071844s" podCreationTimestamp="2025-12-12 06:53:45 +0000 UTC" firstStartedPulling="2025-12-12 06:53:46.674774905 +0000 UTC m=+314.246156174" lastFinishedPulling="2025-12-12 06:53:50.723340887 +0000 UTC m=+318.294722156" observedRunningTime="2025-12-12 06:53:51.734915052 +0000 UTC m=+319.306296321" watchObservedRunningTime="2025-12-12 06:53:51.737071844 +0000 UTC m=+319.308453113" Dec 12 06:53:51 crc kubenswrapper[4867]: I1212 06:53:51.858377 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:51 crc kubenswrapper[4867]: I1212 06:53:51.858413 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:51 crc kubenswrapper[4867]: I1212 06:53:51.927760 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:52 crc kubenswrapper[4867]: I1212 06:53:52.764175 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hszdl" Dec 12 06:53:52 crc kubenswrapper[4867]: I1212 06:53:52.831598 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:52 crc kubenswrapper[4867]: I1212 06:53:52.831666 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:52 crc kubenswrapper[4867]: I1212 06:53:52.875489 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:53 crc kubenswrapper[4867]: I1212 06:53:53.767513 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9dng" Dec 12 06:53:54 crc kubenswrapper[4867]: I1212 06:53:54.223508 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:54 crc kubenswrapper[4867]: I1212 06:53:54.223548 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:54 crc kubenswrapper[4867]: I1212 06:53:54.266216 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:54 crc kubenswrapper[4867]: I1212 06:53:54.793198 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qqqrr" Dec 12 06:53:55 crc kubenswrapper[4867]: I1212 06:53:55.405612 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:55 crc kubenswrapper[4867]: I1212 06:53:55.405916 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:53:56 crc kubenswrapper[4867]: I1212 06:53:56.443059 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2hql5" podUID="196dd20e-43b3-4f25-a61c-a94d958bb555" containerName="registry-server" probeResult="failure" output=< Dec 12 06:53:56 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 06:53:56 crc kubenswrapper[4867]: > Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.523768 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bb7j7"] Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.525026 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.528078 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.528191 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tprl7"] Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.529297 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.529918 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.534865 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.553936 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tprl7"] Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.558749 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bb7j7"] Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644764 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644847 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644872 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpzqd\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-kube-api-access-lpzqd\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644929 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644952 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47rpk\" (UniqueName: \"kubernetes.io/projected/d6e1399d-7725-4ea9-805d-1a363cc0bd09-kube-api-access-47rpk\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644970 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-trusted-ca\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.644993 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.645018 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-bound-sa-token\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.645033 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.645065 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-certificates\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.645087 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-tls\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.713408 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746262 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-bound-sa-token\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746304 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746351 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-certificates\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746384 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-tls\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746411 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746433 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpzqd\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-kube-api-access-lpzqd\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746474 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746490 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47rpk\" (UniqueName: \"kubernetes.io/projected/d6e1399d-7725-4ea9-805d-1a363cc0bd09-kube-api-access-47rpk\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-trusted-ca\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.746521 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.747116 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.748183 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.748420 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-trusted-ca\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.748718 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-certificates\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.751438 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-registry-tls\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.751544 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.751970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d6e1399d-7725-4ea9-805d-1a363cc0bd09-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.776615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47rpk\" (UniqueName: \"kubernetes.io/projected/d6e1399d-7725-4ea9-805d-1a363cc0bd09-kube-api-access-47rpk\") pod \"marketplace-operator-79b997595-bb7j7\" (UID: \"d6e1399d-7725-4ea9-805d-1a363cc0bd09\") " pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.782568 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-bound-sa-token\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.783788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpzqd\" (UniqueName: \"kubernetes.io/projected/16fb70d7-f3aa-45fd-adeb-3de7abe8591b-kube-api-access-lpzqd\") pod \"image-registry-66df7c8f76-tprl7\" (UID: \"16fb70d7-f3aa-45fd-adeb-3de7abe8591b\") " pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.845340 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:00 crc kubenswrapper[4867]: I1212 06:54:00.852444 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.297325 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bb7j7"] Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.380993 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tprl7"] Dec 12 06:54:01 crc kubenswrapper[4867]: W1212 06:54:01.385177 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16fb70d7_f3aa_45fd_adeb_3de7abe8591b.slice/crio-29db84b49b154710b804690377f4d0c946aa08f019255e8e65ecfcba60b7bd23 WatchSource:0}: Error finding container 29db84b49b154710b804690377f4d0c946aa08f019255e8e65ecfcba60b7bd23: Status 404 returned error can't find the container with id 29db84b49b154710b804690377f4d0c946aa08f019255e8e65ecfcba60b7bd23 Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.772668 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" event={"ID":"d6e1399d-7725-4ea9-805d-1a363cc0bd09","Type":"ContainerStarted","Data":"6929889a877ad45a61b4e300c3e958d915e2c3c6efac412bfea5c8133a0c3199"} Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.773045 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" event={"ID":"d6e1399d-7725-4ea9-805d-1a363cc0bd09","Type":"ContainerStarted","Data":"d60230554fc086ad16663025f3affad4c56f48381f3d93479957343ef76c5850"} Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.773065 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.775026 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" event={"ID":"16fb70d7-f3aa-45fd-adeb-3de7abe8591b","Type":"ContainerStarted","Data":"70028c179ddcd0ef0844af613970649f73136f6344ed32ece181c372ad95b177"} Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.775070 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" event={"ID":"16fb70d7-f3aa-45fd-adeb-3de7abe8591b","Type":"ContainerStarted","Data":"29db84b49b154710b804690377f4d0c946aa08f019255e8e65ecfcba60b7bd23"} Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.775151 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.776525 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.788515 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bb7j7" podStartSLOduration=1.788503635 podStartE2EDuration="1.788503635s" podCreationTimestamp="2025-12-12 06:54:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:54:01.785191384 +0000 UTC m=+329.356572653" watchObservedRunningTime="2025-12-12 06:54:01.788503635 +0000 UTC m=+329.359884894" Dec 12 06:54:01 crc kubenswrapper[4867]: I1212 06:54:01.844345 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" podStartSLOduration=1.844329909 podStartE2EDuration="1.844329909s" podCreationTimestamp="2025-12-12 06:54:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:54:01.842117455 +0000 UTC m=+329.413498724" watchObservedRunningTime="2025-12-12 06:54:01.844329909 +0000 UTC m=+329.415711178" Dec 12 06:54:05 crc kubenswrapper[4867]: I1212 06:54:05.441558 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:54:05 crc kubenswrapper[4867]: I1212 06:54:05.491713 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2hql5" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.213862 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.214417 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" podUID="090c2564-d726-4214-b0c3-578018da0b8e" containerName="controller-manager" containerID="cri-o://f51504d9217a0046b2c38c3b8a9a7efe9371cd3925b3b636c11c1d4c7bf2df00" gracePeriod=30 Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.315838 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.316308 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" podUID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" containerName="route-controller-manager" containerID="cri-o://591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6" gracePeriod=30 Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.681836 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.806026 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config\") pod \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.806099 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert\") pod \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.806144 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca\") pod \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.806163 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blzst\" (UniqueName: \"kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst\") pod \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\" (UID: \"07c44bf3-87fc-41fe-b953-291e1b8f8b6a\") " Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.807187 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca" (OuterVolumeSpecName: "client-ca") pod "07c44bf3-87fc-41fe-b953-291e1b8f8b6a" (UID: "07c44bf3-87fc-41fe-b953-291e1b8f8b6a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.807204 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config" (OuterVolumeSpecName: "config") pod "07c44bf3-87fc-41fe-b953-291e1b8f8b6a" (UID: "07c44bf3-87fc-41fe-b953-291e1b8f8b6a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.811619 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst" (OuterVolumeSpecName: "kube-api-access-blzst") pod "07c44bf3-87fc-41fe-b953-291e1b8f8b6a" (UID: "07c44bf3-87fc-41fe-b953-291e1b8f8b6a"). InnerVolumeSpecName "kube-api-access-blzst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.811877 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "07c44bf3-87fc-41fe-b953-291e1b8f8b6a" (UID: "07c44bf3-87fc-41fe-b953-291e1b8f8b6a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.832304 4867 generic.go:334] "Generic (PLEG): container finished" podID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" containerID="591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6" exitCode=0 Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.832371 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" event={"ID":"07c44bf3-87fc-41fe-b953-291e1b8f8b6a","Type":"ContainerDied","Data":"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6"} Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.832392 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.832409 4867 scope.go:117] "RemoveContainer" containerID="591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.832399 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7" event={"ID":"07c44bf3-87fc-41fe-b953-291e1b8f8b6a","Type":"ContainerDied","Data":"c58a1b5120ce3e850c187e603c8e1e32b6488914403ef9b705f1fbc1b059943b"} Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.834115 4867 generic.go:334] "Generic (PLEG): container finished" podID="090c2564-d726-4214-b0c3-578018da0b8e" containerID="f51504d9217a0046b2c38c3b8a9a7efe9371cd3925b3b636c11c1d4c7bf2df00" exitCode=0 Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.834158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" event={"ID":"090c2564-d726-4214-b0c3-578018da0b8e","Type":"ContainerDied","Data":"f51504d9217a0046b2c38c3b8a9a7efe9371cd3925b3b636c11c1d4c7bf2df00"} Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.847463 4867 scope.go:117] "RemoveContainer" containerID="591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6" Dec 12 06:54:11 crc kubenswrapper[4867]: E1212 06:54:11.848213 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6\": container with ID starting with 591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6 not found: ID does not exist" containerID="591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.848270 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6"} err="failed to get container status \"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6\": rpc error: code = NotFound desc = could not find container \"591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6\": container with ID starting with 591c3b2dcabf2fecce80e5d5ee88d9ebe58da67c4d6cd37d8404c41c7b90b4a6 not found: ID does not exist" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.860316 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.864384 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbb74dfd9-dp4l7"] Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.908134 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.908158 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.908167 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:11 crc kubenswrapper[4867]: I1212 06:54:11.909034 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blzst\" (UniqueName: \"kubernetes.io/projected/07c44bf3-87fc-41fe-b953-291e1b8f8b6a-kube-api-access-blzst\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.081044 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.212453 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config\") pod \"090c2564-d726-4214-b0c3-578018da0b8e\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.212513 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl7fl\" (UniqueName: \"kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl\") pod \"090c2564-d726-4214-b0c3-578018da0b8e\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.212610 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles\") pod \"090c2564-d726-4214-b0c3-578018da0b8e\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.212660 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca\") pod \"090c2564-d726-4214-b0c3-578018da0b8e\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.212693 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert\") pod \"090c2564-d726-4214-b0c3-578018da0b8e\" (UID: \"090c2564-d726-4214-b0c3-578018da0b8e\") " Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.213275 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "090c2564-d726-4214-b0c3-578018da0b8e" (UID: "090c2564-d726-4214-b0c3-578018da0b8e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.213499 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca" (OuterVolumeSpecName: "client-ca") pod "090c2564-d726-4214-b0c3-578018da0b8e" (UID: "090c2564-d726-4214-b0c3-578018da0b8e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.213602 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config" (OuterVolumeSpecName: "config") pod "090c2564-d726-4214-b0c3-578018da0b8e" (UID: "090c2564-d726-4214-b0c3-578018da0b8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.216482 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl" (OuterVolumeSpecName: "kube-api-access-xl7fl") pod "090c2564-d726-4214-b0c3-578018da0b8e" (UID: "090c2564-d726-4214-b0c3-578018da0b8e"). InnerVolumeSpecName "kube-api-access-xl7fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.216587 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "090c2564-d726-4214-b0c3-578018da0b8e" (UID: "090c2564-d726-4214-b0c3-578018da0b8e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.314390 4867 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/090c2564-d726-4214-b0c3-578018da0b8e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.314433 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-config\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.314446 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl7fl\" (UniqueName: \"kubernetes.io/projected/090c2564-d726-4214-b0c3-578018da0b8e-kube-api-access-xl7fl\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.314463 4867 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.314474 4867 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/090c2564-d726-4214-b0c3-578018da0b8e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.842507 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.845282 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" path="/var/lib/kubelet/pods/07c44bf3-87fc-41fe-b953-291e1b8f8b6a/volumes" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.846033 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58bf4dd977-7rnkr" event={"ID":"090c2564-d726-4214-b0c3-578018da0b8e","Type":"ContainerDied","Data":"494b670fb6fb6b21bc15a7bcd2c8326f61deb20ee81b32cc0ce331c5d0af0fd8"} Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.846086 4867 scope.go:117] "RemoveContainer" containerID="f51504d9217a0046b2c38c3b8a9a7efe9371cd3925b3b636c11c1d4c7bf2df00" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.882753 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.887061 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-58bf4dd977-7rnkr"] Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.921982 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bccfc4758-hc7ht"] Dec 12 06:54:12 crc kubenswrapper[4867]: E1212 06:54:12.922267 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" containerName="route-controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.922281 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" containerName="route-controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: E1212 06:54:12.922298 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="090c2564-d726-4214-b0c3-578018da0b8e" containerName="controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.922304 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="090c2564-d726-4214-b0c3-578018da0b8e" containerName="controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.922393 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="090c2564-d726-4214-b0c3-578018da0b8e" containerName="controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.922404 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="07c44bf3-87fc-41fe-b953-291e1b8f8b6a" containerName="route-controller-manager" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.922830 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925010 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925238 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925336 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925375 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925423 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.925755 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.932773 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k"] Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.933677 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936163 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936463 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936713 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936775 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936787 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.936874 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.937127 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.946311 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k"] Dec 12 06:54:12 crc kubenswrapper[4867]: I1212 06:54:12.952380 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bccfc4758-hc7ht"] Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.027218 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/170cf073-962b-40f6-bfe7-458cbeb01f4c-serving-cert\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.027676 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nh9l\" (UniqueName: \"kubernetes.io/projected/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-kube-api-access-6nh9l\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.027880 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-config\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.028057 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-config\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.028279 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-client-ca\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.028431 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-proxy-ca-bundles\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.028640 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-client-ca\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.028844 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-serving-cert\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.029055 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks2sb\" (UniqueName: \"kubernetes.io/projected/170cf073-962b-40f6-bfe7-458cbeb01f4c-kube-api-access-ks2sb\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.130457 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-client-ca\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.130731 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-proxy-ca-bundles\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.130867 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-client-ca\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.130982 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-serving-cert\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131113 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks2sb\" (UniqueName: \"kubernetes.io/projected/170cf073-962b-40f6-bfe7-458cbeb01f4c-kube-api-access-ks2sb\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131242 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/170cf073-962b-40f6-bfe7-458cbeb01f4c-serving-cert\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131403 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-client-ca\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131408 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nh9l\" (UniqueName: \"kubernetes.io/projected/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-kube-api-access-6nh9l\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131469 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-config\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131507 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-config\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.131838 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-proxy-ca-bundles\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.132603 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-config\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.133333 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-client-ca\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.133936 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/170cf073-962b-40f6-bfe7-458cbeb01f4c-config\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.134611 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-serving-cert\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.145053 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/170cf073-962b-40f6-bfe7-458cbeb01f4c-serving-cert\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.148617 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nh9l\" (UniqueName: \"kubernetes.io/projected/d9b382d3-e17e-40b1-ae5a-f50bca2e1201-kube-api-access-6nh9l\") pod \"controller-manager-7bccfc4758-hc7ht\" (UID: \"d9b382d3-e17e-40b1-ae5a-f50bca2e1201\") " pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.151051 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks2sb\" (UniqueName: \"kubernetes.io/projected/170cf073-962b-40f6-bfe7-458cbeb01f4c-kube-api-access-ks2sb\") pod \"route-controller-manager-c546f67c8-hgb5k\" (UID: \"170cf073-962b-40f6-bfe7-458cbeb01f4c\") " pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.239061 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.252124 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.645128 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bccfc4758-hc7ht"] Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.678847 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k"] Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.853921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" event={"ID":"170cf073-962b-40f6-bfe7-458cbeb01f4c","Type":"ContainerStarted","Data":"005f39325eff3ce81e3cb601dc9585d983264d0ea2b4a484d9135b500b2ec7aa"} Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.854197 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.854209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" event={"ID":"170cf073-962b-40f6-bfe7-458cbeb01f4c","Type":"ContainerStarted","Data":"33019698dde40995151bdbe52c51326a0cf6ab8364d8eb637bbec3163108402e"} Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.855847 4867 patch_prober.go:28] interesting pod/route-controller-manager-c546f67c8-hgb5k container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.855905 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" podUID="170cf073-962b-40f6-bfe7-458cbeb01f4c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.856468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" event={"ID":"d9b382d3-e17e-40b1-ae5a-f50bca2e1201","Type":"ContainerStarted","Data":"c35a780940801ce1b4af32cf44866ba3e74b840d1d61167f8023f1566da9d65d"} Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.856537 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" event={"ID":"d9b382d3-e17e-40b1-ae5a-f50bca2e1201","Type":"ContainerStarted","Data":"9668429884db2089df8b055a6a0725dabab2b5280568e855295d09126f7d8943"} Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.858139 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.860677 4867 patch_prober.go:28] interesting pod/controller-manager-7bccfc4758-hc7ht container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.860734 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" podUID="d9b382d3-e17e-40b1-ae5a-f50bca2e1201" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.879297 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" podStartSLOduration=2.87927339 podStartE2EDuration="2.87927339s" podCreationTimestamp="2025-12-12 06:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:54:13.874558965 +0000 UTC m=+341.445940254" watchObservedRunningTime="2025-12-12 06:54:13.87927339 +0000 UTC m=+341.450654679" Dec 12 06:54:13 crc kubenswrapper[4867]: I1212 06:54:13.896406 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" podStartSLOduration=2.896387458 podStartE2EDuration="2.896387458s" podCreationTimestamp="2025-12-12 06:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 06:54:13.892994175 +0000 UTC m=+341.464375464" watchObservedRunningTime="2025-12-12 06:54:13.896387458 +0000 UTC m=+341.467768737" Dec 12 06:54:14 crc kubenswrapper[4867]: I1212 06:54:14.849101 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="090c2564-d726-4214-b0c3-578018da0b8e" path="/var/lib/kubelet/pods/090c2564-d726-4214-b0c3-578018da0b8e/volumes" Dec 12 06:54:14 crc kubenswrapper[4867]: I1212 06:54:14.869602 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bccfc4758-hc7ht" Dec 12 06:54:14 crc kubenswrapper[4867]: I1212 06:54:14.870551 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c546f67c8-hgb5k" Dec 12 06:54:20 crc kubenswrapper[4867]: I1212 06:54:20.857684 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tprl7" Dec 12 06:54:20 crc kubenswrapper[4867]: I1212 06:54:20.954311 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:54:28 crc kubenswrapper[4867]: I1212 06:54:28.989186 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:54:28 crc kubenswrapper[4867]: I1212 06:54:28.989778 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.007504 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" podUID="4cc708f7-46a4-4947-8ba9-c960d302a348" containerName="registry" containerID="cri-o://e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34" gracePeriod=30 Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.425080 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.474904 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475049 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svbfg\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475124 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475187 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475410 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475473 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475563 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.475630 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted\") pod \"4cc708f7-46a4-4947-8ba9-c960d302a348\" (UID: \"4cc708f7-46a4-4947-8ba9-c960d302a348\") " Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.476810 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.477218 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.481525 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.482580 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.483769 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg" (OuterVolumeSpecName: "kube-api-access-svbfg") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "kube-api-access-svbfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.486865 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.495058 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.506488 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4cc708f7-46a4-4947-8ba9-c960d302a348" (UID: "4cc708f7-46a4-4947-8ba9-c960d302a348"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576570 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576608 4867 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4cc708f7-46a4-4947-8ba9-c960d302a348-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576620 4867 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576630 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svbfg\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-kube-api-access-svbfg\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576640 4867 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4cc708f7-46a4-4947-8ba9-c960d302a348-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576651 4867 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4cc708f7-46a4-4947-8ba9-c960d302a348-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:46 crc kubenswrapper[4867]: I1212 06:54:46.576659 4867 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cc708f7-46a4-4947-8ba9-c960d302a348-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.055639 4867 generic.go:334] "Generic (PLEG): container finished" podID="4cc708f7-46a4-4947-8ba9-c960d302a348" containerID="e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34" exitCode=0 Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.055692 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" event={"ID":"4cc708f7-46a4-4947-8ba9-c960d302a348","Type":"ContainerDied","Data":"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34"} Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.057373 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" event={"ID":"4cc708f7-46a4-4947-8ba9-c960d302a348","Type":"ContainerDied","Data":"8135db2195c7eb173f3f4a1de78c783ac531745d6ad0576d36535dd81faea807"} Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.055734 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zcz7" Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.057438 4867 scope.go:117] "RemoveContainer" containerID="e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34" Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.084825 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.085794 4867 scope.go:117] "RemoveContainer" containerID="e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34" Dec 12 06:54:47 crc kubenswrapper[4867]: E1212 06:54:47.086355 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34\": container with ID starting with e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34 not found: ID does not exist" containerID="e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34" Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.086384 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34"} err="failed to get container status \"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34\": rpc error: code = NotFound desc = could not find container \"e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34\": container with ID starting with e9da274deee9e5746c39e5daf8d8c67cb5d1a827a58047e873d58ff153d4cf34 not found: ID does not exist" Dec 12 06:54:47 crc kubenswrapper[4867]: I1212 06:54:47.090921 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zcz7"] Dec 12 06:54:48 crc kubenswrapper[4867]: I1212 06:54:48.843817 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc708f7-46a4-4947-8ba9-c960d302a348" path="/var/lib/kubelet/pods/4cc708f7-46a4-4947-8ba9-c960d302a348/volumes" Dec 12 06:54:58 crc kubenswrapper[4867]: I1212 06:54:58.988493 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:54:58 crc kubenswrapper[4867]: I1212 06:54:58.989516 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:55:28 crc kubenswrapper[4867]: I1212 06:55:28.988923 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:55:28 crc kubenswrapper[4867]: I1212 06:55:28.989602 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:55:28 crc kubenswrapper[4867]: I1212 06:55:28.989699 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:55:28 crc kubenswrapper[4867]: I1212 06:55:28.990796 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 06:55:28 crc kubenswrapper[4867]: I1212 06:55:28.990935 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88" gracePeriod=600 Dec 12 06:55:29 crc kubenswrapper[4867]: I1212 06:55:29.293786 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88" exitCode=0 Dec 12 06:55:29 crc kubenswrapper[4867]: I1212 06:55:29.293837 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88"} Dec 12 06:55:29 crc kubenswrapper[4867]: I1212 06:55:29.293908 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973"} Dec 12 06:55:29 crc kubenswrapper[4867]: I1212 06:55:29.293930 4867 scope.go:117] "RemoveContainer" containerID="239367b3411959afdbece0f02169430803adacad70097541bbd53edad55496af" Dec 12 06:57:33 crc kubenswrapper[4867]: I1212 06:57:33.007392 4867 scope.go:117] "RemoveContainer" containerID="a8929b2dc908ead36c3325990a96c8e7e5c73b8ef8d50d2f338edb4b423b272d" Dec 12 06:57:33 crc kubenswrapper[4867]: I1212 06:57:33.025454 4867 scope.go:117] "RemoveContainer" containerID="823aaef7d3679c19382e449d2b9a5a00ad525209d66d1b1310b491103bce4f6b" Dec 12 06:57:58 crc kubenswrapper[4867]: I1212 06:57:58.989122 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:57:58 crc kubenswrapper[4867]: I1212 06:57:58.990089 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:58:28 crc kubenswrapper[4867]: I1212 06:58:28.989214 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:58:28 crc kubenswrapper[4867]: I1212 06:58:28.990315 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:58:58 crc kubenswrapper[4867]: I1212 06:58:58.988557 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 06:58:58 crc kubenswrapper[4867]: I1212 06:58:58.989183 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 06:58:58 crc kubenswrapper[4867]: I1212 06:58:58.989267 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 06:58:58 crc kubenswrapper[4867]: I1212 06:58:58.989966 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 06:58:58 crc kubenswrapper[4867]: I1212 06:58:58.990064 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973" gracePeriod=600 Dec 12 06:58:59 crc kubenswrapper[4867]: I1212 06:58:59.728715 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973" exitCode=0 Dec 12 06:58:59 crc kubenswrapper[4867]: I1212 06:58:59.728763 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973"} Dec 12 06:58:59 crc kubenswrapper[4867]: I1212 06:58:59.729399 4867 scope.go:117] "RemoveContainer" containerID="dc705b310d3970985b2fb0774e75206b28b2f5750adaf1b764f72abba3fb9e88" Dec 12 06:59:00 crc kubenswrapper[4867]: I1212 06:59:00.740493 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14"} Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.166370 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk"] Dec 12 07:00:00 crc kubenswrapper[4867]: E1212 07:00:00.167118 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc708f7-46a4-4947-8ba9-c960d302a348" containerName="registry" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.167135 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc708f7-46a4-4947-8ba9-c960d302a348" containerName="registry" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.167319 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc708f7-46a4-4947-8ba9-c960d302a348" containerName="registry" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.167762 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.171686 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.172134 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.176076 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk"] Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.253943 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59z7q\" (UniqueName: \"kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.254007 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.254033 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.354818 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59z7q\" (UniqueName: \"kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.354870 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.354900 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.355881 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.362372 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.381522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59z7q\" (UniqueName: \"kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q\") pod \"collect-profiles-29425380-d6wpk\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.494775 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:00 crc kubenswrapper[4867]: I1212 07:00:00.688482 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk"] Dec 12 07:00:01 crc kubenswrapper[4867]: I1212 07:00:01.065627 4867 generic.go:334] "Generic (PLEG): container finished" podID="c40f17a4-60eb-4246-aa3f-aa2cf573542d" containerID="86e82217785c1207c428940654715be7d9cee5dd0cfca8e19105f8e3ebbcc836" exitCode=0 Dec 12 07:00:01 crc kubenswrapper[4867]: I1212 07:00:01.065711 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" event={"ID":"c40f17a4-60eb-4246-aa3f-aa2cf573542d","Type":"ContainerDied","Data":"86e82217785c1207c428940654715be7d9cee5dd0cfca8e19105f8e3ebbcc836"} Dec 12 07:00:01 crc kubenswrapper[4867]: I1212 07:00:01.065763 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" event={"ID":"c40f17a4-60eb-4246-aa3f-aa2cf573542d","Type":"ContainerStarted","Data":"77a3cdc60bec2d9b21321167d548be463bf747d5c7a6a23237eb90f49fe87872"} Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.270825 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.385325 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59z7q\" (UniqueName: \"kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q\") pod \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.385374 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume\") pod \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.385408 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume\") pod \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\" (UID: \"c40f17a4-60eb-4246-aa3f-aa2cf573542d\") " Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.386268 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume" (OuterVolumeSpecName: "config-volume") pod "c40f17a4-60eb-4246-aa3f-aa2cf573542d" (UID: "c40f17a4-60eb-4246-aa3f-aa2cf573542d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.390075 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c40f17a4-60eb-4246-aa3f-aa2cf573542d" (UID: "c40f17a4-60eb-4246-aa3f-aa2cf573542d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.391068 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q" (OuterVolumeSpecName: "kube-api-access-59z7q") pod "c40f17a4-60eb-4246-aa3f-aa2cf573542d" (UID: "c40f17a4-60eb-4246-aa3f-aa2cf573542d"). InnerVolumeSpecName "kube-api-access-59z7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.486377 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59z7q\" (UniqueName: \"kubernetes.io/projected/c40f17a4-60eb-4246-aa3f-aa2cf573542d-kube-api-access-59z7q\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.486466 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c40f17a4-60eb-4246-aa3f-aa2cf573542d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:02 crc kubenswrapper[4867]: I1212 07:00:02.486482 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c40f17a4-60eb-4246-aa3f-aa2cf573542d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:03 crc kubenswrapper[4867]: I1212 07:00:03.078610 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" event={"ID":"c40f17a4-60eb-4246-aa3f-aa2cf573542d","Type":"ContainerDied","Data":"77a3cdc60bec2d9b21321167d548be463bf747d5c7a6a23237eb90f49fe87872"} Dec 12 07:00:03 crc kubenswrapper[4867]: I1212 07:00:03.078662 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77a3cdc60bec2d9b21321167d548be463bf747d5c7a6a23237eb90f49fe87872" Dec 12 07:00:03 crc kubenswrapper[4867]: I1212 07:00:03.078721 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.086304 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-kdqtv"] Dec 12 07:00:12 crc kubenswrapper[4867]: E1212 07:00:12.088298 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c40f17a4-60eb-4246-aa3f-aa2cf573542d" containerName="collect-profiles" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.088408 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c40f17a4-60eb-4246-aa3f-aa2cf573542d" containerName="collect-profiles" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.088616 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c40f17a4-60eb-4246-aa3f-aa2cf573542d" containerName="collect-profiles" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.089363 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.092648 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.092676 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.092769 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.092833 4867 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-sdnsw" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.099041 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kdqtv"] Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.100441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.100539 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmqq7\" (UniqueName: \"kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.100587 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.201874 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.202011 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.202086 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmqq7\" (UniqueName: \"kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.202198 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.203322 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.224818 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmqq7\" (UniqueName: \"kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7\") pod \"crc-storage-crc-kdqtv\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.411494 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.581365 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-kdqtv"] Dec 12 07:00:12 crc kubenswrapper[4867]: I1212 07:00:12.598411 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:00:13 crc kubenswrapper[4867]: I1212 07:00:13.130208 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kdqtv" event={"ID":"0587769e-ee2a-40bd-8f22-0d8fbf2f883a","Type":"ContainerStarted","Data":"74103b9709e2a4f3e8aab62c1d8cbcab077d32e4f975f0157b6c99c3c1194bd6"} Dec 12 07:00:16 crc kubenswrapper[4867]: I1212 07:00:16.148154 4867 generic.go:334] "Generic (PLEG): container finished" podID="0587769e-ee2a-40bd-8f22-0d8fbf2f883a" containerID="5e2bb9506306aa9fc906723b6edb756c61d2ea95e9ff94842567322e6385105d" exitCode=0 Dec 12 07:00:16 crc kubenswrapper[4867]: I1212 07:00:16.148213 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kdqtv" event={"ID":"0587769e-ee2a-40bd-8f22-0d8fbf2f883a","Type":"ContainerDied","Data":"5e2bb9506306aa9fc906723b6edb756c61d2ea95e9ff94842567322e6385105d"} Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.341905 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.471539 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt\") pod \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.471686 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmqq7\" (UniqueName: \"kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7\") pod \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.471678 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "0587769e-ee2a-40bd-8f22-0d8fbf2f883a" (UID: "0587769e-ee2a-40bd-8f22-0d8fbf2f883a"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.471921 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage\") pod \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\" (UID: \"0587769e-ee2a-40bd-8f22-0d8fbf2f883a\") " Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.472392 4867 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.476125 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7" (OuterVolumeSpecName: "kube-api-access-zmqq7") pod "0587769e-ee2a-40bd-8f22-0d8fbf2f883a" (UID: "0587769e-ee2a-40bd-8f22-0d8fbf2f883a"). InnerVolumeSpecName "kube-api-access-zmqq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.487055 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "0587769e-ee2a-40bd-8f22-0d8fbf2f883a" (UID: "0587769e-ee2a-40bd-8f22-0d8fbf2f883a"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.573687 4867 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:17 crc kubenswrapper[4867]: I1212 07:00:17.573725 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmqq7\" (UniqueName: \"kubernetes.io/projected/0587769e-ee2a-40bd-8f22-0d8fbf2f883a-kube-api-access-zmqq7\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:18 crc kubenswrapper[4867]: I1212 07:00:18.159808 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-kdqtv" event={"ID":"0587769e-ee2a-40bd-8f22-0d8fbf2f883a","Type":"ContainerDied","Data":"74103b9709e2a4f3e8aab62c1d8cbcab077d32e4f975f0157b6c99c3c1194bd6"} Dec 12 07:00:18 crc kubenswrapper[4867]: I1212 07:00:18.159850 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74103b9709e2a4f3e8aab62c1d8cbcab077d32e4f975f0157b6c99c3c1194bd6" Dec 12 07:00:18 crc kubenswrapper[4867]: I1212 07:00:18.159880 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-kdqtv" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.314065 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbsqz"] Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315003 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-controller" containerID="cri-o://0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315138 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315237 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="sbdb" containerID="cri-o://0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315260 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-node" containerID="cri-o://2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315343 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-acl-logging" containerID="cri-o://a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315139 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="northd" containerID="cri-o://4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.315360 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="nbdb" containerID="cri-o://a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.355529 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" containerID="cri-o://ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" gracePeriod=30 Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.596753 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/3.log" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.599253 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovn-acl-logging/0.log" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.599904 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovn-controller/0.log" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.600358 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647486 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fdg8t"] Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647681 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647694 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647701 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="sbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647707 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="sbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647717 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647723 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647731 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-acl-logging" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647737 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-acl-logging" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647746 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="northd" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647753 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="northd" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647760 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647766 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647776 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kubecfg-setup" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647781 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kubecfg-setup" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647790 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-node" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647796 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-node" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647804 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0587769e-ee2a-40bd-8f22-0d8fbf2f883a" containerName="storage" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647809 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0587769e-ee2a-40bd-8f22-0d8fbf2f883a" containerName="storage" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647817 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="nbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647822 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="nbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647830 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-ovn-metrics" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647836 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-ovn-metrics" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.647844 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647849 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647929 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647938 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="nbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647944 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647951 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-ovn-metrics" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647959 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0587769e-ee2a-40bd-8f22-0d8fbf2f883a" containerName="storage" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647969 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="kube-rbac-proxy-node" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647976 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="northd" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647987 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.647995 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648003 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="sbdb" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648010 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovn-acl-logging" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.648117 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648127 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: E1212 07:00:21.648138 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648146 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648264 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.648281 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerName="ovnkube-controller" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.650331 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.723913 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.723966 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724020 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket" (OuterVolumeSpecName: "log-socket") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724217 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724290 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724313 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724365 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724377 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724407 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724426 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724447 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724407 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash" (OuterVolumeSpecName: "host-slash") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724422 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724441 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log" (OuterVolumeSpecName: "node-log") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724471 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mhrh\" (UniqueName: \"kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724499 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724519 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724538 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724563 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724907 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724454 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724766 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724797 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724831 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724847 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724856 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724962 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.724877 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725059 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725091 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725017 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725092 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725126 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib\") pod \"5f519f9a-7a2d-4193-8c21-ae7840036dae\" (UID: \"5f519f9a-7a2d-4193-8c21-ae7840036dae\") " Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725140 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725140 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725337 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-systemd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725401 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-ovn\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725429 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725480 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-etc-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725508 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-kubelet\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725540 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-node-log\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-script-lib\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725603 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-netd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725626 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725654 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-env-overrides\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725660 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725721 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-config\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725765 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovn-node-metrics-cert\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725839 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fh8m\" (UniqueName: \"kubernetes.io/projected/46f61fd4-f158-4248-b7b6-a883ddfa3204-kube-api-access-8fh8m\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725893 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-var-lib-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725926 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-netns\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725955 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.725983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-bin\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726038 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-slash\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726065 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-systemd-units\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726126 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-log-socket\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726278 4867 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726303 4867 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726318 4867 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726331 4867 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726343 4867 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726354 4867 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-log-socket\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726366 4867 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726379 4867 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726391 4867 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-slash\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726406 4867 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726419 4867 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-node-log\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726429 4867 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726440 4867 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726451 4867 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726462 4867 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726472 4867 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.726483 4867 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.730178 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh" (OuterVolumeSpecName: "kube-api-access-6mhrh") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "kube-api-access-6mhrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.730278 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.736805 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5f519f9a-7a2d-4193-8c21-ae7840036dae" (UID: "5f519f9a-7a2d-4193-8c21-ae7840036dae"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827831 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-systemd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827903 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-ovn\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827924 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827951 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-etc-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827972 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-kubelet\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.827994 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-node-log\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828017 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-script-lib\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828030 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-ovn\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828070 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-etc-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828089 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-netd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828093 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-kubelet\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828071 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-node-log\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828046 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-netd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828181 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828218 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-env-overrides\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828253 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-config\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828279 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovn-node-metrics-cert\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828306 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fh8m\" (UniqueName: \"kubernetes.io/projected/46f61fd4-f158-4248-b7b6-a883ddfa3204-kube-api-access-8fh8m\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828326 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-var-lib-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828348 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-netns\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828367 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828386 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-bin\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828403 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-slash\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828425 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-systemd-units\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-log-socket\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828468 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-systemd\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828030 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828551 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-var-lib-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828559 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828575 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-run-netns\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828596 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-run-openvswitch\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828620 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-cni-bin\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828638 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-host-slash\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828657 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-systemd-units\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828680 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46f61fd4-f158-4248-b7b6-a883ddfa3204-log-socket\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.828777 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-script-lib\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.829143 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-env-overrides\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.829307 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mhrh\" (UniqueName: \"kubernetes.io/projected/5f519f9a-7a2d-4193-8c21-ae7840036dae-kube-api-access-6mhrh\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.829322 4867 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f519f9a-7a2d-4193-8c21-ae7840036dae-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.829331 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f519f9a-7a2d-4193-8c21-ae7840036dae-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.829536 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovnkube-config\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.831693 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46f61fd4-f158-4248-b7b6-a883ddfa3204-ovn-node-metrics-cert\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.844520 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fh8m\" (UniqueName: \"kubernetes.io/projected/46f61fd4-f158-4248-b7b6-a883ddfa3204-kube-api-access-8fh8m\") pod \"ovnkube-node-fdg8t\" (UID: \"46f61fd4-f158-4248-b7b6-a883ddfa3204\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:21 crc kubenswrapper[4867]: I1212 07:00:21.970171 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.178000 4867 generic.go:334] "Generic (PLEG): container finished" podID="46f61fd4-f158-4248-b7b6-a883ddfa3204" containerID="2e08b3e0b01db3ae017d1645ba97d7e8bdaa2c216658550efdb17e4d0675a45f" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.178079 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerDied","Data":"2e08b3e0b01db3ae017d1645ba97d7e8bdaa2c216658550efdb17e4d0675a45f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.178136 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"c21073d91d7fa247542189c89a3a428c36da38bfa4e95bbabc1beaac04cb8625"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.182148 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/2.log" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.183163 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/1.log" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.183213 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1918c53-1a6d-4317-a320-2e301b5a9db7" containerID="fb4548196b76adb7c889aee7392b15bcfde6b7d6613c5b208454e042140f1889" exitCode=2 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.183252 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerDied","Data":"fb4548196b76adb7c889aee7392b15bcfde6b7d6613c5b208454e042140f1889"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.183293 4867 scope.go:117] "RemoveContainer" containerID="9de15ee860c875b630cb3b9fc4a14dd7473adc665a347cb6ce86598f532813e9" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.183718 4867 scope.go:117] "RemoveContainer" containerID="fb4548196b76adb7c889aee7392b15bcfde6b7d6613c5b208454e042140f1889" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.183935 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z8lss_openshift-multus(f1918c53-1a6d-4317-a320-2e301b5a9db7)\"" pod="openshift-multus/multus-z8lss" podUID="f1918c53-1a6d-4317-a320-2e301b5a9db7" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.185128 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovnkube-controller/3.log" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.188133 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovn-acl-logging/0.log" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.189938 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbsqz_5f519f9a-7a2d-4193-8c21-ae7840036dae/ovn-controller/0.log" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190374 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190407 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190416 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190426 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190435 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190445 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" exitCode=0 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190452 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" exitCode=143 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190465 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f519f9a-7a2d-4193-8c21-ae7840036dae" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" exitCode=143 Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190466 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190496 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190507 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190518 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190530 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190538 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190548 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190559 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190565 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190571 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190577 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190583 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190589 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190595 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190600 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190605 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190612 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190620 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190626 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190630 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190636 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190642 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190647 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190653 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190659 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190664 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190669 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190676 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190683 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190689 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190694 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190699 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190705 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190711 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190716 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190721 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190455 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190726 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190940 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.190990 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbsqz" event={"ID":"5f519f9a-7a2d-4193-8c21-ae7840036dae","Type":"ContainerDied","Data":"b1632ebb28e10ceae66726db235e73630a37f4db4503e12ffbdb5a55ca91a407"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191017 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191026 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191033 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191040 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191069 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191081 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191090 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191098 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191108 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.191116 4867 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.218214 4867 scope.go:117] "RemoveContainer" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.247498 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbsqz"] Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.248675 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.255091 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbsqz"] Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.268022 4867 scope.go:117] "RemoveContainer" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.280932 4867 scope.go:117] "RemoveContainer" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.309715 4867 scope.go:117] "RemoveContainer" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.332812 4867 scope.go:117] "RemoveContainer" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.345475 4867 scope.go:117] "RemoveContainer" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.356241 4867 scope.go:117] "RemoveContainer" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.382956 4867 scope.go:117] "RemoveContainer" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.408109 4867 scope.go:117] "RemoveContainer" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.420571 4867 scope.go:117] "RemoveContainer" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.421060 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": container with ID starting with ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44 not found: ID does not exist" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421093 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} err="failed to get container status \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": rpc error: code = NotFound desc = could not find container \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": container with ID starting with ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421111 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.421469 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": container with ID starting with d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09 not found: ID does not exist" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421491 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} err="failed to get container status \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": rpc error: code = NotFound desc = could not find container \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": container with ID starting with d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421504 4867 scope.go:117] "RemoveContainer" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.421726 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": container with ID starting with 0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c not found: ID does not exist" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421744 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} err="failed to get container status \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": rpc error: code = NotFound desc = could not find container \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": container with ID starting with 0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421758 4867 scope.go:117] "RemoveContainer" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.421976 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": container with ID starting with a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2 not found: ID does not exist" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.421996 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} err="failed to get container status \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": rpc error: code = NotFound desc = could not find container \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": container with ID starting with a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422007 4867 scope.go:117] "RemoveContainer" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.422239 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": container with ID starting with 4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec not found: ID does not exist" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422258 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} err="failed to get container status \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": rpc error: code = NotFound desc = could not find container \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": container with ID starting with 4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422271 4867 scope.go:117] "RemoveContainer" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.422471 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": container with ID starting with 0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206 not found: ID does not exist" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422489 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} err="failed to get container status \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": rpc error: code = NotFound desc = could not find container \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": container with ID starting with 0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422500 4867 scope.go:117] "RemoveContainer" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.422676 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": container with ID starting with 2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33 not found: ID does not exist" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422694 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} err="failed to get container status \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": rpc error: code = NotFound desc = could not find container \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": container with ID starting with 2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.422708 4867 scope.go:117] "RemoveContainer" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.423119 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": container with ID starting with a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50 not found: ID does not exist" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423139 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} err="failed to get container status \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": rpc error: code = NotFound desc = could not find container \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": container with ID starting with a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423152 4867 scope.go:117] "RemoveContainer" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.423600 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": container with ID starting with 0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f not found: ID does not exist" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423613 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} err="failed to get container status \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": rpc error: code = NotFound desc = could not find container \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": container with ID starting with 0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423625 4867 scope.go:117] "RemoveContainer" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: E1212 07:00:22.423855 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": container with ID starting with ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462 not found: ID does not exist" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423876 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} err="failed to get container status \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": rpc error: code = NotFound desc = could not find container \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": container with ID starting with ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.423914 4867 scope.go:117] "RemoveContainer" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424166 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} err="failed to get container status \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": rpc error: code = NotFound desc = could not find container \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": container with ID starting with ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424185 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424547 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} err="failed to get container status \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": rpc error: code = NotFound desc = could not find container \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": container with ID starting with d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424601 4867 scope.go:117] "RemoveContainer" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424853 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} err="failed to get container status \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": rpc error: code = NotFound desc = could not find container \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": container with ID starting with 0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.424871 4867 scope.go:117] "RemoveContainer" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.425084 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} err="failed to get container status \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": rpc error: code = NotFound desc = could not find container \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": container with ID starting with a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.425100 4867 scope.go:117] "RemoveContainer" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.425382 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} err="failed to get container status \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": rpc error: code = NotFound desc = could not find container \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": container with ID starting with 4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.425400 4867 scope.go:117] "RemoveContainer" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.425690 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} err="failed to get container status \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": rpc error: code = NotFound desc = could not find container \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": container with ID starting with 0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426300 4867 scope.go:117] "RemoveContainer" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426486 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} err="failed to get container status \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": rpc error: code = NotFound desc = could not find container \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": container with ID starting with 2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426500 4867 scope.go:117] "RemoveContainer" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426709 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} err="failed to get container status \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": rpc error: code = NotFound desc = could not find container \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": container with ID starting with a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426722 4867 scope.go:117] "RemoveContainer" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426963 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} err="failed to get container status \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": rpc error: code = NotFound desc = could not find container \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": container with ID starting with 0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.426981 4867 scope.go:117] "RemoveContainer" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427215 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} err="failed to get container status \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": rpc error: code = NotFound desc = could not find container \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": container with ID starting with ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427250 4867 scope.go:117] "RemoveContainer" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427494 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} err="failed to get container status \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": rpc error: code = NotFound desc = could not find container \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": container with ID starting with ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427516 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427720 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} err="failed to get container status \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": rpc error: code = NotFound desc = could not find container \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": container with ID starting with d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427742 4867 scope.go:117] "RemoveContainer" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427941 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} err="failed to get container status \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": rpc error: code = NotFound desc = could not find container \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": container with ID starting with 0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.427959 4867 scope.go:117] "RemoveContainer" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428151 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} err="failed to get container status \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": rpc error: code = NotFound desc = could not find container \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": container with ID starting with a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428174 4867 scope.go:117] "RemoveContainer" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428399 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} err="failed to get container status \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": rpc error: code = NotFound desc = could not find container \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": container with ID starting with 4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428421 4867 scope.go:117] "RemoveContainer" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428630 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} err="failed to get container status \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": rpc error: code = NotFound desc = could not find container \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": container with ID starting with 0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428656 4867 scope.go:117] "RemoveContainer" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428830 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} err="failed to get container status \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": rpc error: code = NotFound desc = could not find container \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": container with ID starting with 2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.428852 4867 scope.go:117] "RemoveContainer" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429074 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} err="failed to get container status \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": rpc error: code = NotFound desc = could not find container \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": container with ID starting with a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429101 4867 scope.go:117] "RemoveContainer" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429324 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} err="failed to get container status \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": rpc error: code = NotFound desc = could not find container \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": container with ID starting with 0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429349 4867 scope.go:117] "RemoveContainer" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429520 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} err="failed to get container status \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": rpc error: code = NotFound desc = could not find container \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": container with ID starting with ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.429544 4867 scope.go:117] "RemoveContainer" containerID="ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.430188 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44"} err="failed to get container status \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": rpc error: code = NotFound desc = could not find container \"ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44\": container with ID starting with ee1da791c231767f7ae887d05aa12d6747cdb0e6a826f878eeb1531a2eedfc44 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.430255 4867 scope.go:117] "RemoveContainer" containerID="d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.430676 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09"} err="failed to get container status \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": rpc error: code = NotFound desc = could not find container \"d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09\": container with ID starting with d0ec65b5a0b5aa672db91306dd02210558f97c333cddc94dd7d694c717c46f09 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.430709 4867 scope.go:117] "RemoveContainer" containerID="0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431030 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c"} err="failed to get container status \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": rpc error: code = NotFound desc = could not find container \"0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c\": container with ID starting with 0a604f858b8b96f9240060de08aa63b25c7a1597a2b8d7503c41e5d7cfe4147c not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431135 4867 scope.go:117] "RemoveContainer" containerID="a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431506 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2"} err="failed to get container status \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": rpc error: code = NotFound desc = could not find container \"a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2\": container with ID starting with a0c20b4c999d2375b88ffdb63ffa64601f42a4c811141475afaf8e57fe50fbe2 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431531 4867 scope.go:117] "RemoveContainer" containerID="4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431767 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec"} err="failed to get container status \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": rpc error: code = NotFound desc = could not find container \"4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec\": container with ID starting with 4a93758e6a860fa18885ea1a25e65cd1a90c9738ef71cdcfb1c6f1807165e3ec not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.431791 4867 scope.go:117] "RemoveContainer" containerID="0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432020 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206"} err="failed to get container status \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": rpc error: code = NotFound desc = could not find container \"0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206\": container with ID starting with 0ed8d66ccf80679fe0575167270f4dc2829ad260218a4d080e8ed707e70af206 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432127 4867 scope.go:117] "RemoveContainer" containerID="2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432451 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33"} err="failed to get container status \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": rpc error: code = NotFound desc = could not find container \"2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33\": container with ID starting with 2e33c823396cd235358852bf6b99ba019c96ace8e882bd03bad3c99f67505c33 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432487 4867 scope.go:117] "RemoveContainer" containerID="a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432764 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50"} err="failed to get container status \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": rpc error: code = NotFound desc = could not find container \"a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50\": container with ID starting with a9e09a236e9222ada42647fbcb0326d5ffa53f8bfc4d1c4009b23f77d6f2dd50 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.432789 4867 scope.go:117] "RemoveContainer" containerID="0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.433056 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f"} err="failed to get container status \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": rpc error: code = NotFound desc = could not find container \"0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f\": container with ID starting with 0d850bdb026010d95d92149b5dba94d3acf4c078b548863748b93c11211c8f1f not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.433084 4867 scope.go:117] "RemoveContainer" containerID="ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.433363 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462"} err="failed to get container status \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": rpc error: code = NotFound desc = could not find container \"ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462\": container with ID starting with ddf7714b9eb7be105dbdbce680f5201ed31812c16638c5baf0600044b6bcb462 not found: ID does not exist" Dec 12 07:00:22 crc kubenswrapper[4867]: I1212 07:00:22.851019 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f519f9a-7a2d-4193-8c21-ae7840036dae" path="/var/lib/kubelet/pods/5f519f9a-7a2d-4193-8c21-ae7840036dae/volumes" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204140 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"93ba6dcea8113173b9e086a5b2dd4482d2d28a3c7eb154c743d9ac1511142fdc"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"c877e5b3191fa3838decc62060957c7ff0ba25e54f1fafad9b0ba604f88b7cf1"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204434 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"a896aaa4ab5ef7cc578c0ecd42ccd60cdc992fba408a53e79ae7cf569524e7f5"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204446 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"515e87324d83c01aa382b2469de52aa810abe3c59b4f55eccb3e8b92b546fd91"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204454 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"a710122fe0a25cba9c09c17b99fa6d4cc2bff8168877b78a44c5b341f9533c74"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.204466 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"1cdb6d7590acfec6bbcda7e9195a9c48ae2d6743cbedf3eda1a04e03e9fe8a39"} Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.205796 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/2.log" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.792874 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb"] Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.793838 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.795839 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.951570 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.951616 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5vl6\" (UniqueName: \"kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:23 crc kubenswrapper[4867]: I1212 07:00:23.951658 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.052454 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.052724 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5vl6\" (UniqueName: \"kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.052814 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.052897 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.053082 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.076883 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5vl6\" (UniqueName: \"kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: I1212 07:00:24.106182 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: E1212 07:00:24.124041 4867 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(28bd85c278e86902caf18a92b174b0891cbc72f8254fc34ffd915a7de92ae45c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 12 07:00:24 crc kubenswrapper[4867]: E1212 07:00:24.124136 4867 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(28bd85c278e86902caf18a92b174b0891cbc72f8254fc34ffd915a7de92ae45c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: E1212 07:00:24.124168 4867 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(28bd85c278e86902caf18a92b174b0891cbc72f8254fc34ffd915a7de92ae45c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:24 crc kubenswrapper[4867]: E1212 07:00:24.124287 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace(3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace(3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(28bd85c278e86902caf18a92b174b0891cbc72f8254fc34ffd915a7de92ae45c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" Dec 12 07:00:26 crc kubenswrapper[4867]: I1212 07:00:26.222853 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"c22215215c1fbf23a1729939d579f05e1eddd5c51cad94d1851aa49e85913ea7"} Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.199521 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb"] Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.199956 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.200436 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:28 crc kubenswrapper[4867]: E1212 07:00:28.221330 4867 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(f4534b8960c8c69cdb11a73e0573ea5c5ff45a4c1d42c8ca7c67b6a841c17271): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 12 07:00:28 crc kubenswrapper[4867]: E1212 07:00:28.221396 4867 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(f4534b8960c8c69cdb11a73e0573ea5c5ff45a4c1d42c8ca7c67b6a841c17271): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:28 crc kubenswrapper[4867]: E1212 07:00:28.221423 4867 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(f4534b8960c8c69cdb11a73e0573ea5c5ff45a4c1d42c8ca7c67b6a841c17271): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:28 crc kubenswrapper[4867]: E1212 07:00:28.221482 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace(3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace(3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_openshift-marketplace_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08_0(f4534b8960c8c69cdb11a73e0573ea5c5ff45a4c1d42c8ca7c67b6a841c17271): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.235966 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" event={"ID":"46f61fd4-f158-4248-b7b6-a883ddfa3204","Type":"ContainerStarted","Data":"414e89c664822f2d63149b1f1b73e029d388dbe4b687f8b8d23935bf85624bf9"} Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.236278 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.236308 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.260445 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:28 crc kubenswrapper[4867]: I1212 07:00:28.273369 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" podStartSLOduration=7.273354664 podStartE2EDuration="7.273354664s" podCreationTimestamp="2025-12-12 07:00:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:00:28.271927755 +0000 UTC m=+715.843309024" watchObservedRunningTime="2025-12-12 07:00:28.273354664 +0000 UTC m=+715.844735933" Dec 12 07:00:29 crc kubenswrapper[4867]: I1212 07:00:29.240949 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:29 crc kubenswrapper[4867]: I1212 07:00:29.264084 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:37 crc kubenswrapper[4867]: I1212 07:00:37.838355 4867 scope.go:117] "RemoveContainer" containerID="fb4548196b76adb7c889aee7392b15bcfde6b7d6613c5b208454e042140f1889" Dec 12 07:00:38 crc kubenswrapper[4867]: I1212 07:00:38.285726 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z8lss_f1918c53-1a6d-4317-a320-2e301b5a9db7/kube-multus/2.log" Dec 12 07:00:38 crc kubenswrapper[4867]: I1212 07:00:38.286070 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z8lss" event={"ID":"f1918c53-1a6d-4317-a320-2e301b5a9db7","Type":"ContainerStarted","Data":"f07ff940765122284d9369389467fae67ed29a077221346354780a1e4055b410"} Dec 12 07:00:40 crc kubenswrapper[4867]: I1212 07:00:40.837487 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:40 crc kubenswrapper[4867]: I1212 07:00:40.838294 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:41 crc kubenswrapper[4867]: I1212 07:00:41.240348 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb"] Dec 12 07:00:41 crc kubenswrapper[4867]: W1212 07:00:41.246364 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3951e0b1_ac70_4fe4_b9ec_2c33ce2bba08.slice/crio-d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123 WatchSource:0}: Error finding container d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123: Status 404 returned error can't find the container with id d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123 Dec 12 07:00:41 crc kubenswrapper[4867]: I1212 07:00:41.309418 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerStarted","Data":"d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123"} Dec 12 07:00:42 crc kubenswrapper[4867]: I1212 07:00:42.316551 4867 generic.go:334] "Generic (PLEG): container finished" podID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerID="f42ddefbf8e17921bee697b97c8b6754b7519da8b9360c26064a50e39a96206f" exitCode=0 Dec 12 07:00:42 crc kubenswrapper[4867]: I1212 07:00:42.316637 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerDied","Data":"f42ddefbf8e17921bee697b97c8b6754b7519da8b9360c26064a50e39a96206f"} Dec 12 07:00:44 crc kubenswrapper[4867]: I1212 07:00:44.330309 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerStarted","Data":"0a75d2fb1eda10bcd2e7aec4f119b390f5a2f54757ca34476442efd98f905319"} Dec 12 07:00:45 crc kubenswrapper[4867]: I1212 07:00:45.336765 4867 generic.go:334] "Generic (PLEG): container finished" podID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerID="0a75d2fb1eda10bcd2e7aec4f119b390f5a2f54757ca34476442efd98f905319" exitCode=0 Dec 12 07:00:45 crc kubenswrapper[4867]: I1212 07:00:45.336826 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerDied","Data":"0a75d2fb1eda10bcd2e7aec4f119b390f5a2f54757ca34476442efd98f905319"} Dec 12 07:00:47 crc kubenswrapper[4867]: I1212 07:00:47.351171 4867 generic.go:334] "Generic (PLEG): container finished" podID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerID="2da7fed85b6596016942d460c970737c6951e281a19a6faf0d7c81f5a506d1ce" exitCode=0 Dec 12 07:00:47 crc kubenswrapper[4867]: I1212 07:00:47.351262 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerDied","Data":"2da7fed85b6596016942d460c970737c6951e281a19a6faf0d7c81f5a506d1ce"} Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.738997 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.867715 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle\") pod \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.868043 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util\") pod \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.868129 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5vl6\" (UniqueName: \"kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6\") pod \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\" (UID: \"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08\") " Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.868448 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle" (OuterVolumeSpecName: "bundle") pod "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" (UID: "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.878697 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util" (OuterVolumeSpecName: "util") pod "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" (UID: "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.879418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6" (OuterVolumeSpecName: "kube-api-access-v5vl6") pod "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" (UID: "3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08"). InnerVolumeSpecName "kube-api-access-v5vl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.969729 4867 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-util\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.969772 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5vl6\" (UniqueName: \"kubernetes.io/projected/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-kube-api-access-v5vl6\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:48 crc kubenswrapper[4867]: I1212 07:00:48.969794 4867 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:00:49 crc kubenswrapper[4867]: I1212 07:00:49.365774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" event={"ID":"3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08","Type":"ContainerDied","Data":"d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123"} Dec 12 07:00:49 crc kubenswrapper[4867]: I1212 07:00:49.365831 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4e6f79f697e3853e97f6a2cf6630cdd628616956be622a53f47cfd76daf9123" Dec 12 07:00:49 crc kubenswrapper[4867]: I1212 07:00:49.365897 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb" Dec 12 07:00:51 crc kubenswrapper[4867]: I1212 07:00:51.991834 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdg8t" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.471022 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-ts9zq"] Dec 12 07:00:55 crc kubenswrapper[4867]: E1212 07:00:55.471745 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="pull" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.471756 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="pull" Dec 12 07:00:55 crc kubenswrapper[4867]: E1212 07:00:55.471767 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="util" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.471773 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="util" Dec 12 07:00:55 crc kubenswrapper[4867]: E1212 07:00:55.471784 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="extract" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.471790 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="extract" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.471888 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08" containerName="extract" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.472217 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.475118 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.477755 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-7t8hm" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.481457 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.492066 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-ts9zq"] Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.545636 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vzz9\" (UniqueName: \"kubernetes.io/projected/01d83fb8-4fb3-43f4-a597-daed3e63f0bb-kube-api-access-9vzz9\") pod \"nmstate-operator-6769fb99d-ts9zq\" (UID: \"01d83fb8-4fb3-43f4-a597-daed3e63f0bb\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.646704 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vzz9\" (UniqueName: \"kubernetes.io/projected/01d83fb8-4fb3-43f4-a597-daed3e63f0bb-kube-api-access-9vzz9\") pod \"nmstate-operator-6769fb99d-ts9zq\" (UID: \"01d83fb8-4fb3-43f4-a597-daed3e63f0bb\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.667210 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vzz9\" (UniqueName: \"kubernetes.io/projected/01d83fb8-4fb3-43f4-a597-daed3e63f0bb-kube-api-access-9vzz9\") pod \"nmstate-operator-6769fb99d-ts9zq\" (UID: \"01d83fb8-4fb3-43f4-a597-daed3e63f0bb\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.794908 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" Dec 12 07:00:55 crc kubenswrapper[4867]: I1212 07:00:55.967150 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-ts9zq"] Dec 12 07:00:56 crc kubenswrapper[4867]: I1212 07:00:56.401112 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" event={"ID":"01d83fb8-4fb3-43f4-a597-daed3e63f0bb","Type":"ContainerStarted","Data":"10d5e2df163dd4e882118be7ed895c559737099ba47511ec550d7b1ebaf1afb9"} Dec 12 07:00:58 crc kubenswrapper[4867]: I1212 07:00:58.410648 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" event={"ID":"01d83fb8-4fb3-43f4-a597-daed3e63f0bb","Type":"ContainerStarted","Data":"f67b13a702e0b4dd2dacc0e2880c82b3fe5471bcad8f25b9c754926c83fd0bdf"} Dec 12 07:00:58 crc kubenswrapper[4867]: I1212 07:00:58.427834 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-ts9zq" podStartSLOduration=1.169140957 podStartE2EDuration="3.427815771s" podCreationTimestamp="2025-12-12 07:00:55 +0000 UTC" firstStartedPulling="2025-12-12 07:00:55.981262694 +0000 UTC m=+743.552643963" lastFinishedPulling="2025-12-12 07:00:58.239937508 +0000 UTC m=+745.811318777" observedRunningTime="2025-12-12 07:00:58.426370855 +0000 UTC m=+745.997752134" watchObservedRunningTime="2025-12-12 07:00:58.427815771 +0000 UTC m=+745.999197040" Dec 12 07:01:02 crc kubenswrapper[4867]: I1212 07:01:02.753707 4867 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.193011 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.194002 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.197895 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-trwjm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.203256 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-c7b78"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.204386 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.205881 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.240562 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-wzr2t"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.242090 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.268436 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-c7b78"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.296246 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.339045 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.339854 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.341646 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jtqzb" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.349145 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.349381 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350068 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350739 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-ovs-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350792 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-dbus-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350863 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r88k7\" (UniqueName: \"kubernetes.io/projected/f4cebb09-194f-4459-b31e-43fdbd632cc1-kube-api-access-r88k7\") pod \"nmstate-metrics-7f7f7578db-2jnfh\" (UID: \"f4cebb09-194f-4459-b31e-43fdbd632cc1\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350900 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thv9v\" (UniqueName: \"kubernetes.io/projected/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-kube-api-access-thv9v\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350924 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpf9d\" (UniqueName: \"kubernetes.io/projected/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-kube-api-access-mpf9d\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350945 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.350967 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-nmstate-lock\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ce76a66-a552-43f5-b37e-d62306ce2422-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452537 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r88k7\" (UniqueName: \"kubernetes.io/projected/f4cebb09-194f-4459-b31e-43fdbd632cc1-kube-api-access-r88k7\") pod \"nmstate-metrics-7f7f7578db-2jnfh\" (UID: \"f4cebb09-194f-4459-b31e-43fdbd632cc1\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452607 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thv9v\" (UniqueName: \"kubernetes.io/projected/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-kube-api-access-thv9v\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452631 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452651 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpf9d\" (UniqueName: \"kubernetes.io/projected/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-kube-api-access-mpf9d\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452691 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-nmstate-lock\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452713 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-ovs-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452740 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-dbus-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452781 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt8gc\" (UniqueName: \"kubernetes.io/projected/1ce76a66-a552-43f5-b37e-d62306ce2422-kube-api-access-vt8gc\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.452820 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ce76a66-a552-43f5-b37e-d62306ce2422-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.454421 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-ovs-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.454510 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-nmstate-lock\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.454650 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-dbus-socket\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.459544 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.483307 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r88k7\" (UniqueName: \"kubernetes.io/projected/f4cebb09-194f-4459-b31e-43fdbd632cc1-kube-api-access-r88k7\") pod \"nmstate-metrics-7f7f7578db-2jnfh\" (UID: \"f4cebb09-194f-4459-b31e-43fdbd632cc1\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.485142 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpf9d\" (UniqueName: \"kubernetes.io/projected/ed16d134-4ec5-44bc-bb0b-5281bea6dad3-kube-api-access-mpf9d\") pod \"nmstate-handler-wzr2t\" (UID: \"ed16d134-4ec5-44bc-bb0b-5281bea6dad3\") " pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.485757 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thv9v\" (UniqueName: \"kubernetes.io/projected/444fb8e2-d658-4b0f-b572-d6acbfb93fcd-kube-api-access-thv9v\") pod \"nmstate-webhook-f8fb84555-c7b78\" (UID: \"444fb8e2-d658-4b0f-b572-d6acbfb93fcd\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.518417 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-94b79ccc9-kxwdk"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.519212 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.540872 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-94b79ccc9-kxwdk"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.553590 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt8gc\" (UniqueName: \"kubernetes.io/projected/1ce76a66-a552-43f5-b37e-d62306ce2422-kube-api-access-vt8gc\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.553734 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ce76a66-a552-43f5-b37e-d62306ce2422-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.553756 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ce76a66-a552-43f5-b37e-d62306ce2422-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.554771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ce76a66-a552-43f5-b37e-d62306ce2422-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.556760 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ce76a66-a552-43f5-b37e-d62306ce2422-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.567702 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.573833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt8gc\" (UniqueName: \"kubernetes.io/projected/1ce76a66-a552-43f5-b37e-d62306ce2422-kube-api-access-vt8gc\") pod \"nmstate-console-plugin-6ff7998486-22djm\" (UID: \"1ce76a66-a552-43f5-b37e-d62306ce2422\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.579599 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.592646 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.657921 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvzf8\" (UniqueName: \"kubernetes.io/projected/7b540333-9ad6-4b35-8717-52df746f7d04-kube-api-access-wvzf8\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658204 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-service-ca\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658274 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-oauth-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658301 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-console-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658343 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-oauth-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.658359 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-trusted-ca-bundle\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.664997 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759208 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-oauth-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759267 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-console-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759301 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-oauth-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759317 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-trusted-ca-bundle\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759352 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvzf8\" (UniqueName: \"kubernetes.io/projected/7b540333-9ad6-4b35-8717-52df746f7d04-kube-api-access-wvzf8\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759371 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.759393 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-service-ca\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.760013 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh"] Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.760292 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-service-ca\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.760474 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-oauth-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.761074 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-console-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.761366 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b540333-9ad6-4b35-8717-52df746f7d04-trusted-ca-bundle\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.765609 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-serving-cert\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.766429 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b540333-9ad6-4b35-8717-52df746f7d04-console-oauth-config\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.778740 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvzf8\" (UniqueName: \"kubernetes.io/projected/7b540333-9ad6-4b35-8717-52df746f7d04-kube-api-access-wvzf8\") pod \"console-94b79ccc9-kxwdk\" (UID: \"7b540333-9ad6-4b35-8717-52df746f7d04\") " pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: W1212 07:01:04.780683 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4cebb09_194f_4459_b31e_43fdbd632cc1.slice/crio-4397a4fbe82d9041ab2ec795efa8fc8d714aa9b46cb48e2b92ccb9faed7d7eeb WatchSource:0}: Error finding container 4397a4fbe82d9041ab2ec795efa8fc8d714aa9b46cb48e2b92ccb9faed7d7eeb: Status 404 returned error can't find the container with id 4397a4fbe82d9041ab2ec795efa8fc8d714aa9b46cb48e2b92ccb9faed7d7eeb Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.841068 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:04 crc kubenswrapper[4867]: I1212 07:01:04.863722 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm"] Dec 12 07:01:04 crc kubenswrapper[4867]: W1212 07:01:04.867713 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ce76a66_a552_43f5_b37e_d62306ce2422.slice/crio-c1de410b923328998be1dfbfae3df551ca77bf7f2deb88bec9b2379c84af935b WatchSource:0}: Error finding container c1de410b923328998be1dfbfae3df551ca77bf7f2deb88bec9b2379c84af935b: Status 404 returned error can't find the container with id c1de410b923328998be1dfbfae3df551ca77bf7f2deb88bec9b2379c84af935b Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.001033 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-94b79ccc9-kxwdk"] Dec 12 07:01:05 crc kubenswrapper[4867]: W1212 07:01:05.005970 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b540333_9ad6_4b35_8717_52df746f7d04.slice/crio-0b2ee31bf9228959230458aea04709ef93106a786271dda14bd82fb7a5ee3e05 WatchSource:0}: Error finding container 0b2ee31bf9228959230458aea04709ef93106a786271dda14bd82fb7a5ee3e05: Status 404 returned error can't find the container with id 0b2ee31bf9228959230458aea04709ef93106a786271dda14bd82fb7a5ee3e05 Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.028073 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-c7b78"] Dec 12 07:01:05 crc kubenswrapper[4867]: W1212 07:01:05.030753 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod444fb8e2_d658_4b0f_b572_d6acbfb93fcd.slice/crio-cb5d1d8dbed2cf042638ca02e9bb740f5efd9739bea10e5f4980123e15590021 WatchSource:0}: Error finding container cb5d1d8dbed2cf042638ca02e9bb740f5efd9739bea10e5f4980123e15590021: Status 404 returned error can't find the container with id cb5d1d8dbed2cf042638ca02e9bb740f5efd9739bea10e5f4980123e15590021 Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.443588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" event={"ID":"444fb8e2-d658-4b0f-b572-d6acbfb93fcd","Type":"ContainerStarted","Data":"cb5d1d8dbed2cf042638ca02e9bb740f5efd9739bea10e5f4980123e15590021"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.445814 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" event={"ID":"1ce76a66-a552-43f5-b37e-d62306ce2422","Type":"ContainerStarted","Data":"c1de410b923328998be1dfbfae3df551ca77bf7f2deb88bec9b2379c84af935b"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.447009 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wzr2t" event={"ID":"ed16d134-4ec5-44bc-bb0b-5281bea6dad3","Type":"ContainerStarted","Data":"cbdd220dd910958ac6c57a60f5cf8d43f3f7319795416fd856814d7f09bd927d"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.447924 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" event={"ID":"f4cebb09-194f-4459-b31e-43fdbd632cc1","Type":"ContainerStarted","Data":"4397a4fbe82d9041ab2ec795efa8fc8d714aa9b46cb48e2b92ccb9faed7d7eeb"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.449340 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-94b79ccc9-kxwdk" event={"ID":"7b540333-9ad6-4b35-8717-52df746f7d04","Type":"ContainerStarted","Data":"b0464730a92819b121c23fe74442948ef70bfce77f11bdf006dd6650df342cca"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.449392 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-94b79ccc9-kxwdk" event={"ID":"7b540333-9ad6-4b35-8717-52df746f7d04","Type":"ContainerStarted","Data":"0b2ee31bf9228959230458aea04709ef93106a786271dda14bd82fb7a5ee3e05"} Dec 12 07:01:05 crc kubenswrapper[4867]: I1212 07:01:05.472088 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-94b79ccc9-kxwdk" podStartSLOduration=1.472068888 podStartE2EDuration="1.472068888s" podCreationTimestamp="2025-12-12 07:01:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:01:05.468416816 +0000 UTC m=+753.039798095" watchObservedRunningTime="2025-12-12 07:01:05.472068888 +0000 UTC m=+753.043450157" Dec 12 07:01:07 crc kubenswrapper[4867]: I1212 07:01:07.463546 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" event={"ID":"444fb8e2-d658-4b0f-b572-d6acbfb93fcd","Type":"ContainerStarted","Data":"b70c705eacfa0db06e09c45254e851725caf4eca46047ba42b56abf51bd539f7"} Dec 12 07:01:07 crc kubenswrapper[4867]: I1212 07:01:07.464295 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:07 crc kubenswrapper[4867]: I1212 07:01:07.489277 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" podStartSLOduration=1.217453735 podStartE2EDuration="3.489249897s" podCreationTimestamp="2025-12-12 07:01:04 +0000 UTC" firstStartedPulling="2025-12-12 07:01:05.032503036 +0000 UTC m=+752.603884305" lastFinishedPulling="2025-12-12 07:01:07.304299198 +0000 UTC m=+754.875680467" observedRunningTime="2025-12-12 07:01:07.482046277 +0000 UTC m=+755.053427546" watchObservedRunningTime="2025-12-12 07:01:07.489249897 +0000 UTC m=+755.060631186" Dec 12 07:01:08 crc kubenswrapper[4867]: I1212 07:01:08.469998 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" event={"ID":"1ce76a66-a552-43f5-b37e-d62306ce2422","Type":"ContainerStarted","Data":"c87b8ce0522d5f4c4bbdedbb28698c8c8168e74a9a9963d39528db5f9db185d5"} Dec 12 07:01:08 crc kubenswrapper[4867]: I1212 07:01:08.471362 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-wzr2t" event={"ID":"ed16d134-4ec5-44bc-bb0b-5281bea6dad3","Type":"ContainerStarted","Data":"90aa7f72d018f6a1718d9fbcaf646294ccb57e659a26d18f6404c667bcde44ad"} Dec 12 07:01:08 crc kubenswrapper[4867]: I1212 07:01:08.489701 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-22djm" podStartSLOduration=1.2008963910000001 podStartE2EDuration="4.489678928s" podCreationTimestamp="2025-12-12 07:01:04 +0000 UTC" firstStartedPulling="2025-12-12 07:01:04.87008007 +0000 UTC m=+752.441461339" lastFinishedPulling="2025-12-12 07:01:08.158862607 +0000 UTC m=+755.730243876" observedRunningTime="2025-12-12 07:01:08.485139604 +0000 UTC m=+756.056520873" watchObservedRunningTime="2025-12-12 07:01:08.489678928 +0000 UTC m=+756.061060197" Dec 12 07:01:08 crc kubenswrapper[4867]: I1212 07:01:08.505522 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-wzr2t" podStartSLOduration=1.814628601 podStartE2EDuration="4.505503944s" podCreationTimestamp="2025-12-12 07:01:04 +0000 UTC" firstStartedPulling="2025-12-12 07:01:04.616336058 +0000 UTC m=+752.187717327" lastFinishedPulling="2025-12-12 07:01:07.307211391 +0000 UTC m=+754.878592670" observedRunningTime="2025-12-12 07:01:08.503959646 +0000 UTC m=+756.075340915" watchObservedRunningTime="2025-12-12 07:01:08.505503944 +0000 UTC m=+756.076885213" Dec 12 07:01:09 crc kubenswrapper[4867]: I1212 07:01:09.476790 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:10 crc kubenswrapper[4867]: I1212 07:01:10.484812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" event={"ID":"f4cebb09-194f-4459-b31e-43fdbd632cc1","Type":"ContainerStarted","Data":"7f0dc0191a80c25d9984964581fff5afb0fe17053da20d2d4785125030a2a2f5"} Dec 12 07:01:12 crc kubenswrapper[4867]: I1212 07:01:12.499056 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" event={"ID":"f4cebb09-194f-4459-b31e-43fdbd632cc1","Type":"ContainerStarted","Data":"3561854fca49729980bf2c340e45ff7fbc59fe21334ed413946e37c56e9b0948"} Dec 12 07:01:12 crc kubenswrapper[4867]: I1212 07:01:12.516691 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-2jnfh" podStartSLOduration=1.724876315 podStartE2EDuration="8.516669603s" podCreationTimestamp="2025-12-12 07:01:04 +0000 UTC" firstStartedPulling="2025-12-12 07:01:04.783109823 +0000 UTC m=+752.354491092" lastFinishedPulling="2025-12-12 07:01:11.574903111 +0000 UTC m=+759.146284380" observedRunningTime="2025-12-12 07:01:12.516489479 +0000 UTC m=+760.087870768" watchObservedRunningTime="2025-12-12 07:01:12.516669603 +0000 UTC m=+760.088050872" Dec 12 07:01:14 crc kubenswrapper[4867]: I1212 07:01:14.626685 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-wzr2t" Dec 12 07:01:14 crc kubenswrapper[4867]: I1212 07:01:14.845677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:14 crc kubenswrapper[4867]: I1212 07:01:14.845955 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:14 crc kubenswrapper[4867]: I1212 07:01:14.847009 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:15 crc kubenswrapper[4867]: I1212 07:01:15.518348 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-94b79ccc9-kxwdk" Dec 12 07:01:15 crc kubenswrapper[4867]: I1212 07:01:15.578102 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 07:01:24 crc kubenswrapper[4867]: I1212 07:01:24.585822 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-c7b78" Dec 12 07:01:28 crc kubenswrapper[4867]: I1212 07:01:28.991497 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:01:28 crc kubenswrapper[4867]: I1212 07:01:28.991769 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.663259 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6"] Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.665467 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.667556 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.674321 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6"] Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.811996 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb8fc\" (UniqueName: \"kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.812405 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.812487 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.913994 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.914053 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb8fc\" (UniqueName: \"kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.914122 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.914779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.914833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:38 crc kubenswrapper[4867]: I1212 07:01:38.936279 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb8fc\" (UniqueName: \"kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:39 crc kubenswrapper[4867]: I1212 07:01:39.021710 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:39 crc kubenswrapper[4867]: I1212 07:01:39.399588 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6"] Dec 12 07:01:39 crc kubenswrapper[4867]: I1212 07:01:39.643120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" event={"ID":"ce7963a0-6daf-4995-b782-f6d03ee9f1e7","Type":"ContainerStarted","Data":"1ee2a15ded5175745a877765316c8801ff517791fac8b10a2e47585a08fa8d2f"} Dec 12 07:01:40 crc kubenswrapper[4867]: I1212 07:01:40.620058 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-6xxwt" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" containerID="cri-o://dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137" gracePeriod=15 Dec 12 07:01:40 crc kubenswrapper[4867]: I1212 07:01:40.650046 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerID="3ba48801027f901188dda20e59778198e461f2afda175698fca537e28523c293" exitCode=0 Dec 12 07:01:40 crc kubenswrapper[4867]: I1212 07:01:40.650131 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" event={"ID":"ce7963a0-6daf-4995-b782-f6d03ee9f1e7","Type":"ContainerDied","Data":"3ba48801027f901188dda20e59778198e461f2afda175698fca537e28523c293"} Dec 12 07:01:40 crc kubenswrapper[4867]: I1212 07:01:40.977825 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6xxwt_759c6366-d32d-47d9-89c4-ce7d274637c5/console/0.log" Dec 12 07:01:40 crc kubenswrapper[4867]: I1212 07:01:40.978141 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139705 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139756 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139824 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139862 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139903 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxft9\" (UniqueName: \"kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139924 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.139982 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert\") pod \"759c6366-d32d-47d9-89c4-ce7d274637c5\" (UID: \"759c6366-d32d-47d9-89c4-ce7d274637c5\") " Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.140903 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.140913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.140947 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config" (OuterVolumeSpecName: "console-config") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.140940 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca" (OuterVolumeSpecName: "service-ca") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.145907 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9" (OuterVolumeSpecName: "kube-api-access-gxft9") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "kube-api-access-gxft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.146016 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.146416 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "759c6366-d32d-47d9-89c4-ce7d274637c5" (UID: "759c6366-d32d-47d9-89c4-ce7d274637c5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241417 4867 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241457 4867 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241467 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxft9\" (UniqueName: \"kubernetes.io/projected/759c6366-d32d-47d9-89c4-ce7d274637c5-kube-api-access-gxft9\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241478 4867 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-console-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241486 4867 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/759c6366-d32d-47d9-89c4-ce7d274637c5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241497 4867 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.241505 4867 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/759c6366-d32d-47d9-89c4-ce7d274637c5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657651 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6xxwt_759c6366-d32d-47d9-89c4-ce7d274637c5/console/0.log" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657704 4867 generic.go:334] "Generic (PLEG): container finished" podID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerID="dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137" exitCode=2 Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6xxwt" event={"ID":"759c6366-d32d-47d9-89c4-ce7d274637c5","Type":"ContainerDied","Data":"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137"} Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657758 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6xxwt" event={"ID":"759c6366-d32d-47d9-89c4-ce7d274637c5","Type":"ContainerDied","Data":"80245644461f06327bd42133e8da9b5841bf7dac272aa948a8abb04a5526d616"} Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657774 4867 scope.go:117] "RemoveContainer" containerID="dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.657786 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6xxwt" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.680031 4867 scope.go:117] "RemoveContainer" containerID="dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137" Dec 12 07:01:41 crc kubenswrapper[4867]: E1212 07:01:41.680601 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137\": container with ID starting with dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137 not found: ID does not exist" containerID="dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.680634 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137"} err="failed to get container status \"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137\": rpc error: code = NotFound desc = could not find container \"dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137\": container with ID starting with dc762e91aa3f3a5d5c33d272cc6f895e64a868650a5df79b2213ebb450f3e137 not found: ID does not exist" Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.692099 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 07:01:41 crc kubenswrapper[4867]: I1212 07:01:41.695655 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-6xxwt"] Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.221868 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:42 crc kubenswrapper[4867]: E1212 07:01:42.222636 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.222730 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.222905 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" containerName="console" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.223782 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.231311 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.355501 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.355554 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctbwk\" (UniqueName: \"kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.355653 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.456369 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.456443 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.456458 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctbwk\" (UniqueName: \"kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.457133 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.457369 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.473909 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctbwk\" (UniqueName: \"kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk\") pod \"redhat-operators-ljgb7\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.564850 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.678729 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerID="5f2167963093fca6b5a8aef5c8e27481be3181545961537eb45bd35613dbd21b" exitCode=0 Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.679074 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" event={"ID":"ce7963a0-6daf-4995-b782-f6d03ee9f1e7","Type":"ContainerDied","Data":"5f2167963093fca6b5a8aef5c8e27481be3181545961537eb45bd35613dbd21b"} Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.849823 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="759c6366-d32d-47d9-89c4-ce7d274637c5" path="/var/lib/kubelet/pods/759c6366-d32d-47d9-89c4-ce7d274637c5/volumes" Dec 12 07:01:42 crc kubenswrapper[4867]: I1212 07:01:42.951708 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:43 crc kubenswrapper[4867]: I1212 07:01:43.691194 4867 generic.go:334] "Generic (PLEG): container finished" podID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerID="8683a5f23e5f581efe531e1916471313e67a098401a949ad2a6e9116c4daba40" exitCode=0 Dec 12 07:01:43 crc kubenswrapper[4867]: I1212 07:01:43.691277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerDied","Data":"8683a5f23e5f581efe531e1916471313e67a098401a949ad2a6e9116c4daba40"} Dec 12 07:01:43 crc kubenswrapper[4867]: I1212 07:01:43.691604 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerStarted","Data":"9597ed785732487a4a570566f830b0cb18512e3f623433de48318522da42dadb"} Dec 12 07:01:43 crc kubenswrapper[4867]: I1212 07:01:43.693854 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerID="e8398b61a861a7b1b1cd2916cc26c0f4c87f00a3daf63bac37488cba19daba50" exitCode=0 Dec 12 07:01:43 crc kubenswrapper[4867]: I1212 07:01:43.693890 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" event={"ID":"ce7963a0-6daf-4995-b782-f6d03ee9f1e7","Type":"ContainerDied","Data":"e8398b61a861a7b1b1cd2916cc26c0f4c87f00a3daf63bac37488cba19daba50"} Dec 12 07:01:44 crc kubenswrapper[4867]: I1212 07:01:44.700966 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerStarted","Data":"f356ed1da7c3ae6a5c0fe02267ac3e6100c5899c140fa7bb098b76569f70ef76"} Dec 12 07:01:44 crc kubenswrapper[4867]: I1212 07:01:44.908982 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.085348 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util\") pod \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.085498 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle\") pod \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.085522 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb8fc\" (UniqueName: \"kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc\") pod \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\" (UID: \"ce7963a0-6daf-4995-b782-f6d03ee9f1e7\") " Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.086418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle" (OuterVolumeSpecName: "bundle") pod "ce7963a0-6daf-4995-b782-f6d03ee9f1e7" (UID: "ce7963a0-6daf-4995-b782-f6d03ee9f1e7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.091371 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc" (OuterVolumeSpecName: "kube-api-access-fb8fc") pod "ce7963a0-6daf-4995-b782-f6d03ee9f1e7" (UID: "ce7963a0-6daf-4995-b782-f6d03ee9f1e7"). InnerVolumeSpecName "kube-api-access-fb8fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.103187 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util" (OuterVolumeSpecName: "util") pod "ce7963a0-6daf-4995-b782-f6d03ee9f1e7" (UID: "ce7963a0-6daf-4995-b782-f6d03ee9f1e7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.186918 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb8fc\" (UniqueName: \"kubernetes.io/projected/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-kube-api-access-fb8fc\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.186952 4867 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.186963 4867 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ce7963a0-6daf-4995-b782-f6d03ee9f1e7-util\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.707056 4867 generic.go:334] "Generic (PLEG): container finished" podID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerID="f356ed1da7c3ae6a5c0fe02267ac3e6100c5899c140fa7bb098b76569f70ef76" exitCode=0 Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.707125 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerDied","Data":"f356ed1da7c3ae6a5c0fe02267ac3e6100c5899c140fa7bb098b76569f70ef76"} Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.710560 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" event={"ID":"ce7963a0-6daf-4995-b782-f6d03ee9f1e7","Type":"ContainerDied","Data":"1ee2a15ded5175745a877765316c8801ff517791fac8b10a2e47585a08fa8d2f"} Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.710599 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee2a15ded5175745a877765316c8801ff517791fac8b10a2e47585a08fa8d2f" Dec 12 07:01:45 crc kubenswrapper[4867]: I1212 07:01:45.710603 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6" Dec 12 07:01:46 crc kubenswrapper[4867]: I1212 07:01:46.728366 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerStarted","Data":"140764f20c1b0238ba3947e5fbe30b10a30cb54a3e4658d565dc9aad2aec6390"} Dec 12 07:01:46 crc kubenswrapper[4867]: I1212 07:01:46.746284 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ljgb7" podStartSLOduration=2.219315325 podStartE2EDuration="4.746267964s" podCreationTimestamp="2025-12-12 07:01:42 +0000 UTC" firstStartedPulling="2025-12-12 07:01:43.692999561 +0000 UTC m=+791.264380840" lastFinishedPulling="2025-12-12 07:01:46.21995221 +0000 UTC m=+793.791333479" observedRunningTime="2025-12-12 07:01:46.745661989 +0000 UTC m=+794.317043288" watchObservedRunningTime="2025-12-12 07:01:46.746267964 +0000 UTC m=+794.317649223" Dec 12 07:01:52 crc kubenswrapper[4867]: I1212 07:01:52.565196 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:52 crc kubenswrapper[4867]: I1212 07:01:52.565739 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:52 crc kubenswrapper[4867]: I1212 07:01:52.603978 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:52 crc kubenswrapper[4867]: I1212 07:01:52.796785 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:56 crc kubenswrapper[4867]: I1212 07:01:56.016100 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:56 crc kubenswrapper[4867]: I1212 07:01:56.016570 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ljgb7" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="registry-server" containerID="cri-o://140764f20c1b0238ba3947e5fbe30b10a30cb54a3e4658d565dc9aad2aec6390" gracePeriod=2 Dec 12 07:01:58 crc kubenswrapper[4867]: I1212 07:01:58.790237 4867 generic.go:334] "Generic (PLEG): container finished" podID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerID="140764f20c1b0238ba3947e5fbe30b10a30cb54a3e4658d565dc9aad2aec6390" exitCode=0 Dec 12 07:01:58 crc kubenswrapper[4867]: I1212 07:01:58.790337 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerDied","Data":"140764f20c1b0238ba3947e5fbe30b10a30cb54a3e4658d565dc9aad2aec6390"} Dec 12 07:01:58 crc kubenswrapper[4867]: I1212 07:01:58.988566 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:01:58 crc kubenswrapper[4867]: I1212 07:01:58.988625 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.239311 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247177 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm"] Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247379 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="pull" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247393 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="pull" Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247408 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="util" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247416 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="util" Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247429 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="extract-content" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247437 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="extract-content" Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247449 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="extract" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247455 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="extract" Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247465 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="extract-utilities" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247470 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="extract-utilities" Dec 12 07:01:59 crc kubenswrapper[4867]: E1212 07:01:59.247489 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="registry-server" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247496 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="registry-server" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247593 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" containerName="registry-server" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247609 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7963a0-6daf-4995-b782-f6d03ee9f1e7" containerName="extract" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.247996 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.250616 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-n4wd5" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.250635 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.250859 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.252578 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.252581 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.259025 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl2kl\" (UniqueName: \"kubernetes.io/projected/2bdc1624-a89c-4703-9d6b-12dd126ca87f-kube-api-access-hl2kl\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.259069 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-webhook-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.259141 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-apiservice-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.268000 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.359806 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctbwk\" (UniqueName: \"kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk\") pod \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.359874 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content\") pod \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.360141 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities\") pod \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\" (UID: \"370c28d1-25e6-4f4b-9dc0-c00037acb7ce\") " Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.360327 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-apiservice-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.360357 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl2kl\" (UniqueName: \"kubernetes.io/projected/2bdc1624-a89c-4703-9d6b-12dd126ca87f-kube-api-access-hl2kl\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.360382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-webhook-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.364740 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk" (OuterVolumeSpecName: "kube-api-access-ctbwk") pod "370c28d1-25e6-4f4b-9dc0-c00037acb7ce" (UID: "370c28d1-25e6-4f4b-9dc0-c00037acb7ce"). InnerVolumeSpecName "kube-api-access-ctbwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.365159 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-webhook-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.365465 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities" (OuterVolumeSpecName: "utilities") pod "370c28d1-25e6-4f4b-9dc0-c00037acb7ce" (UID: "370c28d1-25e6-4f4b-9dc0-c00037acb7ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.368778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2bdc1624-a89c-4703-9d6b-12dd126ca87f-apiservice-cert\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.388901 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl2kl\" (UniqueName: \"kubernetes.io/projected/2bdc1624-a89c-4703-9d6b-12dd126ca87f-kube-api-access-hl2kl\") pod \"metallb-operator-controller-manager-85cd4f79db-fjxlm\" (UID: \"2bdc1624-a89c-4703-9d6b-12dd126ca87f\") " pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.461894 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.462134 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctbwk\" (UniqueName: \"kubernetes.io/projected/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-kube-api-access-ctbwk\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.467920 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.468723 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.470885 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.470910 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.471081 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tsk8f" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.484673 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "370c28d1-25e6-4f4b-9dc0-c00037acb7ce" (UID: "370c28d1-25e6-4f4b-9dc0-c00037acb7ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.486081 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.561543 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.563058 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmnwn\" (UniqueName: \"kubernetes.io/projected/2736a731-8ce4-4551-9fec-190c75ef0f04-kube-api-access-tmnwn\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.563115 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-webhook-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.563141 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-apiservice-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.563214 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/370c28d1-25e6-4f4b-9dc0-c00037acb7ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.664601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmnwn\" (UniqueName: \"kubernetes.io/projected/2736a731-8ce4-4551-9fec-190c75ef0f04-kube-api-access-tmnwn\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.665093 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-webhook-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.665117 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-apiservice-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.670682 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-apiservice-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.670982 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2736a731-8ce4-4551-9fec-190c75ef0f04-webhook-cert\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.683277 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmnwn\" (UniqueName: \"kubernetes.io/projected/2736a731-8ce4-4551-9fec-190c75ef0f04-kube-api-access-tmnwn\") pod \"metallb-operator-webhook-server-97cbb5d69-6gxhd\" (UID: \"2736a731-8ce4-4551-9fec-190c75ef0f04\") " pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.783127 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.797913 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ljgb7" event={"ID":"370c28d1-25e6-4f4b-9dc0-c00037acb7ce","Type":"ContainerDied","Data":"9597ed785732487a4a570566f830b0cb18512e3f623433de48318522da42dadb"} Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.797963 4867 scope.go:117] "RemoveContainer" containerID="140764f20c1b0238ba3947e5fbe30b10a30cb54a3e4658d565dc9aad2aec6390" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.798075 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ljgb7" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.806251 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.834604 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.837532 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ljgb7"] Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.837885 4867 scope.go:117] "RemoveContainer" containerID="f356ed1da7c3ae6a5c0fe02267ac3e6100c5899c140fa7bb098b76569f70ef76" Dec 12 07:01:59 crc kubenswrapper[4867]: I1212 07:01:59.853814 4867 scope.go:117] "RemoveContainer" containerID="8683a5f23e5f581efe531e1916471313e67a098401a949ad2a6e9116c4daba40" Dec 12 07:02:00 crc kubenswrapper[4867]: I1212 07:02:00.000723 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd"] Dec 12 07:02:00 crc kubenswrapper[4867]: W1212 07:02:00.005736 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2736a731_8ce4_4551_9fec_190c75ef0f04.slice/crio-efe8d0b75729e80960ba6a4d8be57d3fcfbf362416c959f0ece69e9bed5fd462 WatchSource:0}: Error finding container efe8d0b75729e80960ba6a4d8be57d3fcfbf362416c959f0ece69e9bed5fd462: Status 404 returned error can't find the container with id efe8d0b75729e80960ba6a4d8be57d3fcfbf362416c959f0ece69e9bed5fd462 Dec 12 07:02:00 crc kubenswrapper[4867]: I1212 07:02:00.804966 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" event={"ID":"2bdc1624-a89c-4703-9d6b-12dd126ca87f","Type":"ContainerStarted","Data":"81cd40af266ebe093aae0ca170353db84f93dd757c9713ff482e6e03e0c7aa1c"} Dec 12 07:02:00 crc kubenswrapper[4867]: I1212 07:02:00.806191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" event={"ID":"2736a731-8ce4-4551-9fec-190c75ef0f04","Type":"ContainerStarted","Data":"efe8d0b75729e80960ba6a4d8be57d3fcfbf362416c959f0ece69e9bed5fd462"} Dec 12 07:02:00 crc kubenswrapper[4867]: I1212 07:02:00.845072 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="370c28d1-25e6-4f4b-9dc0-c00037acb7ce" path="/var/lib/kubelet/pods/370c28d1-25e6-4f4b-9dc0-c00037acb7ce/volumes" Dec 12 07:02:03 crc kubenswrapper[4867]: I1212 07:02:03.824165 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" event={"ID":"2bdc1624-a89c-4703-9d6b-12dd126ca87f","Type":"ContainerStarted","Data":"005ff8078b35ab4df3cc36266bd97210028330886c681a6272801bf9a2745dfd"} Dec 12 07:02:03 crc kubenswrapper[4867]: I1212 07:02:03.824713 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:02:03 crc kubenswrapper[4867]: I1212 07:02:03.841076 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" podStartSLOduration=1.4600779990000001 podStartE2EDuration="4.841057754s" podCreationTimestamp="2025-12-12 07:01:59 +0000 UTC" firstStartedPulling="2025-12-12 07:01:59.838314387 +0000 UTC m=+807.409695656" lastFinishedPulling="2025-12-12 07:02:03.219294142 +0000 UTC m=+810.790675411" observedRunningTime="2025-12-12 07:02:03.840934711 +0000 UTC m=+811.412315980" watchObservedRunningTime="2025-12-12 07:02:03.841057754 +0000 UTC m=+811.412439013" Dec 12 07:02:06 crc kubenswrapper[4867]: I1212 07:02:06.844833 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" event={"ID":"2736a731-8ce4-4551-9fec-190c75ef0f04","Type":"ContainerStarted","Data":"58f3d171a1799176852d63ac0d7335e9179bfb53bd98544723fc358149213d90"} Dec 12 07:02:06 crc kubenswrapper[4867]: I1212 07:02:06.845387 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:02:06 crc kubenswrapper[4867]: I1212 07:02:06.862143 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" podStartSLOduration=1.413115013 podStartE2EDuration="7.862127661s" podCreationTimestamp="2025-12-12 07:01:59 +0000 UTC" firstStartedPulling="2025-12-12 07:02:00.008640379 +0000 UTC m=+807.580021648" lastFinishedPulling="2025-12-12 07:02:06.457653037 +0000 UTC m=+814.029034296" observedRunningTime="2025-12-12 07:02:06.860434929 +0000 UTC m=+814.431816198" watchObservedRunningTime="2025-12-12 07:02:06.862127661 +0000 UTC m=+814.433508930" Dec 12 07:02:19 crc kubenswrapper[4867]: I1212 07:02:19.787165 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-97cbb5d69-6gxhd" Dec 12 07:02:28 crc kubenswrapper[4867]: I1212 07:02:28.988769 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:02:28 crc kubenswrapper[4867]: I1212 07:02:28.989363 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:02:28 crc kubenswrapper[4867]: I1212 07:02:28.989412 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:02:28 crc kubenswrapper[4867]: I1212 07:02:28.989981 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:02:28 crc kubenswrapper[4867]: I1212 07:02:28.990057 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14" gracePeriod=600 Dec 12 07:02:29 crc kubenswrapper[4867]: I1212 07:02:29.970891 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14" exitCode=0 Dec 12 07:02:29 crc kubenswrapper[4867]: I1212 07:02:29.970997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14"} Dec 12 07:02:29 crc kubenswrapper[4867]: I1212 07:02:29.971535 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713"} Dec 12 07:02:29 crc kubenswrapper[4867]: I1212 07:02:29.971577 4867 scope.go:117] "RemoveContainer" containerID="036f767dc76ccdae41dd4336557c48513aba8d6ac15a7ee58fd66eaf4c4aa973" Dec 12 07:02:39 crc kubenswrapper[4867]: I1212 07:02:39.564459 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-85cd4f79db-fjxlm" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.205997 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.206922 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.210838 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.213865 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.213910 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbl4\" (UniqueName: \"kubernetes.io/projected/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-kube-api-access-4lbl4\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.215045 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-4rx5r"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.217186 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.217806 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6ldv7" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.219275 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.219834 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.220065 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.303841 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-cwrkp"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.304777 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.307170 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.308117 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-778dz" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.314636 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vvc\" (UniqueName: \"kubernetes.io/projected/68b218be-4b27-4caa-8f64-af253c7a2dff-kube-api-access-56vvc\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.314705 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/68b218be-4b27-4caa-8f64-af253c7a2dff-metallb-excludel2\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.314727 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-conf\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.314745 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-metrics\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.314876 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315015 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.315168 4867 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.315246 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert podName:dbbdf628-904d-4b3a-b07f-40f46b5a57b2 nodeName:}" failed. No retries permitted until 2025-12-12 07:02:40.815209275 +0000 UTC m=+848.386590544 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert") pod "frr-k8s-webhook-server-7784b6fcf-fbwh6" (UID: "dbbdf628-904d-4b3a-b07f-40f46b5a57b2") : secret "frr-k8s-webhook-server-cert" not found Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315545 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-reloader\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315619 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a86044-479c-496a-bce8-8a8b61199127-metrics-certs\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315722 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbl4\" (UniqueName: \"kubernetes.io/projected/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-kube-api-access-4lbl4\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315761 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a86044-479c-496a-bce8-8a8b61199127-frr-startup\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315895 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-sockets\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.315992 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc9zc\" (UniqueName: \"kubernetes.io/projected/45a86044-479c-496a-bce8-8a8b61199127-kube-api-access-vc9zc\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.317011 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.317020 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.326153 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-9wsvj"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.327547 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.334820 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.362054 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-9wsvj"] Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.369363 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbl4\" (UniqueName: \"kubernetes.io/projected/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-kube-api-access-4lbl4\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.416895 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/68b218be-4b27-4caa-8f64-af253c7a2dff-metallb-excludel2\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.416954 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-conf\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.416981 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-cert\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417003 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-metrics\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417024 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417057 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm7cx\" (UniqueName: \"kubernetes.io/projected/d21a5490-c13d-4182-a608-81bfa11f5327-kube-api-access-pm7cx\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417101 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-reloader\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417122 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417142 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a86044-479c-496a-bce8-8a8b61199127-metrics-certs\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417164 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a86044-479c-496a-bce8-8a8b61199127-frr-startup\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417210 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-sockets\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417256 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417287 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc9zc\" (UniqueName: \"kubernetes.io/projected/45a86044-479c-496a-bce8-8a8b61199127-kube-api-access-vc9zc\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.417311 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vvc\" (UniqueName: \"kubernetes.io/projected/68b218be-4b27-4caa-8f64-af253c7a2dff-kube-api-access-56vvc\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.418421 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/68b218be-4b27-4caa-8f64-af253c7a2dff-metallb-excludel2\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.418514 4867 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.418563 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist podName:68b218be-4b27-4caa-8f64-af253c7a2dff nodeName:}" failed. No retries permitted until 2025-12-12 07:02:40.918548612 +0000 UTC m=+848.489929881 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist") pod "speaker-cwrkp" (UID: "68b218be-4b27-4caa-8f64-af253c7a2dff") : secret "metallb-memberlist" not found Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.418790 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-conf\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.418928 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-metrics\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.419026 4867 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.419064 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs podName:68b218be-4b27-4caa-8f64-af253c7a2dff nodeName:}" failed. No retries permitted until 2025-12-12 07:02:40.919053825 +0000 UTC m=+848.490435234 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs") pod "speaker-cwrkp" (UID: "68b218be-4b27-4caa-8f64-af253c7a2dff") : secret "speaker-certs-secret" not found Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.419309 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-reloader\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.419545 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a86044-479c-496a-bce8-8a8b61199127-frr-sockets\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.420706 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a86044-479c-496a-bce8-8a8b61199127-frr-startup\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.442722 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a86044-479c-496a-bce8-8a8b61199127-metrics-certs\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.444527 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vvc\" (UniqueName: \"kubernetes.io/projected/68b218be-4b27-4caa-8f64-af253c7a2dff-kube-api-access-56vvc\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.456330 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc9zc\" (UniqueName: \"kubernetes.io/projected/45a86044-479c-496a-bce8-8a8b61199127-kube-api-access-vc9zc\") pod \"frr-k8s-4rx5r\" (UID: \"45a86044-479c-496a-bce8-8a8b61199127\") " pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.518836 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.518913 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-cert\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.518946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm7cx\" (UniqueName: \"kubernetes.io/projected/d21a5490-c13d-4182-a608-81bfa11f5327-kube-api-access-pm7cx\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.518983 4867 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.519053 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs podName:d21a5490-c13d-4182-a608-81bfa11f5327 nodeName:}" failed. No retries permitted until 2025-12-12 07:02:41.019034848 +0000 UTC m=+848.590416117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs") pod "controller-5bddd4b946-9wsvj" (UID: "d21a5490-c13d-4182-a608-81bfa11f5327") : secret "controller-certs-secret" not found Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.520682 4867 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.532429 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-cert\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.533885 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.536856 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm7cx\" (UniqueName: \"kubernetes.io/projected/d21a5490-c13d-4182-a608-81bfa11f5327-kube-api-access-pm7cx\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.821773 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.825799 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbbdf628-904d-4b3a-b07f-40f46b5a57b2-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-fbwh6\" (UID: \"dbbdf628-904d-4b3a-b07f-40f46b5a57b2\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.825970 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.922957 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.923017 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.924997 4867 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 12 07:02:40 crc kubenswrapper[4867]: E1212 07:02:40.925075 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist podName:68b218be-4b27-4caa-8f64-af253c7a2dff nodeName:}" failed. No retries permitted until 2025-12-12 07:02:41.925057964 +0000 UTC m=+849.496439233 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist") pod "speaker-cwrkp" (UID: "68b218be-4b27-4caa-8f64-af253c7a2dff") : secret "metallb-memberlist" not found Dec 12 07:02:40 crc kubenswrapper[4867]: I1212 07:02:40.929083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-metrics-certs\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.024410 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.027131 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d21a5490-c13d-4182-a608-81bfa11f5327-metrics-certs\") pod \"controller-5bddd4b946-9wsvj\" (UID: \"d21a5490-c13d-4182-a608-81bfa11f5327\") " pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.223829 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6"] Dec 12 07:02:41 crc kubenswrapper[4867]: W1212 07:02:41.230734 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbbdf628_904d_4b3a_b07f_40f46b5a57b2.slice/crio-3acf49e14cd2e15ee35b86ec09c39c3a7824b765ca5ab043b028ab3a583579e3 WatchSource:0}: Error finding container 3acf49e14cd2e15ee35b86ec09c39c3a7824b765ca5ab043b028ab3a583579e3: Status 404 returned error can't find the container with id 3acf49e14cd2e15ee35b86ec09c39c3a7824b765ca5ab043b028ab3a583579e3 Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.241528 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.638751 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-9wsvj"] Dec 12 07:02:41 crc kubenswrapper[4867]: W1212 07:02:41.642206 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd21a5490_c13d_4182_a608_81bfa11f5327.slice/crio-273b47c3ef26b624db6a68dcb8bdd4f71fcb183341d73900bce63c799805f8b0 WatchSource:0}: Error finding container 273b47c3ef26b624db6a68dcb8bdd4f71fcb183341d73900bce63c799805f8b0: Status 404 returned error can't find the container with id 273b47c3ef26b624db6a68dcb8bdd4f71fcb183341d73900bce63c799805f8b0 Dec 12 07:02:41 crc kubenswrapper[4867]: I1212 07:02:41.933654 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:41 crc kubenswrapper[4867]: E1212 07:02:41.933829 4867 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 12 07:02:41 crc kubenswrapper[4867]: E1212 07:02:41.934049 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist podName:68b218be-4b27-4caa-8f64-af253c7a2dff nodeName:}" failed. No retries permitted until 2025-12-12 07:02:43.934031459 +0000 UTC m=+851.505412728 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist") pod "speaker-cwrkp" (UID: "68b218be-4b27-4caa-8f64-af253c7a2dff") : secret "metallb-memberlist" not found Dec 12 07:02:42 crc kubenswrapper[4867]: I1212 07:02:42.057326 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-9wsvj" event={"ID":"d21a5490-c13d-4182-a608-81bfa11f5327","Type":"ContainerStarted","Data":"273b47c3ef26b624db6a68dcb8bdd4f71fcb183341d73900bce63c799805f8b0"} Dec 12 07:02:42 crc kubenswrapper[4867]: I1212 07:02:42.058588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" event={"ID":"dbbdf628-904d-4b3a-b07f-40f46b5a57b2","Type":"ContainerStarted","Data":"3acf49e14cd2e15ee35b86ec09c39c3a7824b765ca5ab043b028ab3a583579e3"} Dec 12 07:02:43 crc kubenswrapper[4867]: I1212 07:02:43.959961 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:43 crc kubenswrapper[4867]: I1212 07:02:43.967393 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/68b218be-4b27-4caa-8f64-af253c7a2dff-memberlist\") pod \"speaker-cwrkp\" (UID: \"68b218be-4b27-4caa-8f64-af253c7a2dff\") " pod="metallb-system/speaker-cwrkp" Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.076189 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-9wsvj" event={"ID":"d21a5490-c13d-4182-a608-81bfa11f5327","Type":"ContainerStarted","Data":"0fa88310436696b16326809b4b4654a8c15025584abbef2c89cb71e44c2e2487"} Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.076252 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-9wsvj" event={"ID":"d21a5490-c13d-4182-a608-81bfa11f5327","Type":"ContainerStarted","Data":"fac4ded0a63712dbcac2f874f3e60382bdaebbc59eceb68098b3bc292614a804"} Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.077126 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.078953 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"0b80c2d9e0319a77bbc45dc891b54a6cef64aa704f367afd7dc63746a7c05e58"} Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.096362 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-9wsvj" podStartSLOduration=4.096341574 podStartE2EDuration="4.096341574s" podCreationTimestamp="2025-12-12 07:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:02:44.094471368 +0000 UTC m=+851.665852657" watchObservedRunningTime="2025-12-12 07:02:44.096341574 +0000 UTC m=+851.667722843" Dec 12 07:02:44 crc kubenswrapper[4867]: I1212 07:02:44.218412 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-cwrkp" Dec 12 07:02:44 crc kubenswrapper[4867]: W1212 07:02:44.255899 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b218be_4b27_4caa_8f64_af253c7a2dff.slice/crio-fd7703ac679d782c8b1461a40db2ee684f04ed4ae9688216a24d978deb444a91 WatchSource:0}: Error finding container fd7703ac679d782c8b1461a40db2ee684f04ed4ae9688216a24d978deb444a91: Status 404 returned error can't find the container with id fd7703ac679d782c8b1461a40db2ee684f04ed4ae9688216a24d978deb444a91 Dec 12 07:02:45 crc kubenswrapper[4867]: I1212 07:02:45.091387 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwrkp" event={"ID":"68b218be-4b27-4caa-8f64-af253c7a2dff","Type":"ContainerStarted","Data":"364daef05e52b99a3e50692705a400c60c8775c6279d082b476233d419773b3a"} Dec 12 07:02:45 crc kubenswrapper[4867]: I1212 07:02:45.091717 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwrkp" event={"ID":"68b218be-4b27-4caa-8f64-af253c7a2dff","Type":"ContainerStarted","Data":"f057d6ae65a21eeedd4630512dac0c48b8e12699f9c5537479a0b0cda9f9e6b4"} Dec 12 07:02:45 crc kubenswrapper[4867]: I1212 07:02:45.091738 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-cwrkp" event={"ID":"68b218be-4b27-4caa-8f64-af253c7a2dff","Type":"ContainerStarted","Data":"fd7703ac679d782c8b1461a40db2ee684f04ed4ae9688216a24d978deb444a91"} Dec 12 07:02:45 crc kubenswrapper[4867]: I1212 07:02:45.091912 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-cwrkp" Dec 12 07:02:45 crc kubenswrapper[4867]: I1212 07:02:45.116594 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-cwrkp" podStartSLOduration=5.116571948 podStartE2EDuration="5.116571948s" podCreationTimestamp="2025-12-12 07:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:02:45.108876087 +0000 UTC m=+852.680257356" watchObservedRunningTime="2025-12-12 07:02:45.116571948 +0000 UTC m=+852.687953217" Dec 12 07:02:49 crc kubenswrapper[4867]: I1212 07:02:49.119413 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" event={"ID":"dbbdf628-904d-4b3a-b07f-40f46b5a57b2","Type":"ContainerStarted","Data":"8ae6c161ecfefe3c279baa1bc5cd29ef49be7414e687585bbcfdb4ff6d8429a5"} Dec 12 07:02:49 crc kubenswrapper[4867]: I1212 07:02:49.120207 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:02:49 crc kubenswrapper[4867]: I1212 07:02:49.121003 4867 generic.go:334] "Generic (PLEG): container finished" podID="45a86044-479c-496a-bce8-8a8b61199127" containerID="6730b9c310b6d92b0615595d607b86a714282b6636829c111639dc41aee886be" exitCode=0 Dec 12 07:02:49 crc kubenswrapper[4867]: I1212 07:02:49.121058 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerDied","Data":"6730b9c310b6d92b0615595d607b86a714282b6636829c111639dc41aee886be"} Dec 12 07:02:49 crc kubenswrapper[4867]: I1212 07:02:49.139205 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" podStartSLOduration=1.568676173 podStartE2EDuration="9.139185825s" podCreationTimestamp="2025-12-12 07:02:40 +0000 UTC" firstStartedPulling="2025-12-12 07:02:41.232706357 +0000 UTC m=+848.804087626" lastFinishedPulling="2025-12-12 07:02:48.803216009 +0000 UTC m=+856.374597278" observedRunningTime="2025-12-12 07:02:49.132650602 +0000 UTC m=+856.704031871" watchObservedRunningTime="2025-12-12 07:02:49.139185825 +0000 UTC m=+856.710567094" Dec 12 07:02:50 crc kubenswrapper[4867]: I1212 07:02:50.128278 4867 generic.go:334] "Generic (PLEG): container finished" podID="45a86044-479c-496a-bce8-8a8b61199127" containerID="ee38961d89d14b82fd2b4ff0460bcde261d3c1ad5804b0e223dd970fe3bce5d8" exitCode=0 Dec 12 07:02:50 crc kubenswrapper[4867]: I1212 07:02:50.128324 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerDied","Data":"ee38961d89d14b82fd2b4ff0460bcde261d3c1ad5804b0e223dd970fe3bce5d8"} Dec 12 07:02:51 crc kubenswrapper[4867]: I1212 07:02:51.134886 4867 generic.go:334] "Generic (PLEG): container finished" podID="45a86044-479c-496a-bce8-8a8b61199127" containerID="f85dd5d67ef624d2690b0e441560d2c5545f9afbbf63b21760b0fc76c88ac4bb" exitCode=0 Dec 12 07:02:51 crc kubenswrapper[4867]: I1212 07:02:51.134949 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerDied","Data":"f85dd5d67ef624d2690b0e441560d2c5545f9afbbf63b21760b0fc76c88ac4bb"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.144843 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"ed6eefbe9847cf39fd00cad3ddc75860256752f18a96b69cf22ec770b5107837"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145179 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"68e82bdca6aa33e625ddd90f77c8a9ee2392c92d8770fb8b175ef41c487b3e77"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145204 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145219 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"53e876962628508e5aef90bea31abca8b26c7a87764236aa99e3ea4ca4c50581"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145253 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"f1e2c337c39db77657e7d890c0cc127c61f4bef71acb491ff8afdd32e5eaf9fb"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145265 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"0cbdb44b32c9d213c3fb103e6836fd0860972796442c0fbae9b5102b897d0d8c"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.145277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-4rx5r" event={"ID":"45a86044-479c-496a-bce8-8a8b61199127","Type":"ContainerStarted","Data":"c9feb01e1b9f3608c99af95b41c4fe02cac9498a513aa57fe926af9a91e5cd77"} Dec 12 07:02:52 crc kubenswrapper[4867]: I1212 07:02:52.175056 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-4rx5r" podStartSLOduration=7.005806228 podStartE2EDuration="12.175027258s" podCreationTimestamp="2025-12-12 07:02:40 +0000 UTC" firstStartedPulling="2025-12-12 07:02:43.660284652 +0000 UTC m=+851.231665921" lastFinishedPulling="2025-12-12 07:02:48.829505672 +0000 UTC m=+856.400886951" observedRunningTime="2025-12-12 07:02:52.171355038 +0000 UTC m=+859.742736307" watchObservedRunningTime="2025-12-12 07:02:52.175027258 +0000 UTC m=+859.746408527" Dec 12 07:02:54 crc kubenswrapper[4867]: I1212 07:02:54.224371 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-cwrkp" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.535710 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.562324 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55"] Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.564076 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.566450 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.591163 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55"] Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.619623 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.628105 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.628166 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.628202 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchh9\" (UniqueName: \"kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.730050 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.730098 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.730133 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchh9\" (UniqueName: \"kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.730590 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.730952 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.749432 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchh9\" (UniqueName: \"kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:55 crc kubenswrapper[4867]: I1212 07:02:55.880717 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:02:56 crc kubenswrapper[4867]: I1212 07:02:56.296573 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55"] Dec 12 07:02:57 crc kubenswrapper[4867]: I1212 07:02:57.174505 4867 generic.go:334] "Generic (PLEG): container finished" podID="b966f81e-701c-49f5-bedf-732815bc91bb" containerID="edfbd212003a400d95bc8e663d44b6aaadc85ec3e9eef1da07458bcb4897e911" exitCode=0 Dec 12 07:02:57 crc kubenswrapper[4867]: I1212 07:02:57.174558 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerDied","Data":"edfbd212003a400d95bc8e663d44b6aaadc85ec3e9eef1da07458bcb4897e911"} Dec 12 07:02:57 crc kubenswrapper[4867]: I1212 07:02:57.174822 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerStarted","Data":"917de5ad1718fe63719eb4494869c25416c107875ff8e9292d59e5eddfa41371"} Dec 12 07:03:00 crc kubenswrapper[4867]: I1212 07:03:00.198609 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerStarted","Data":"aded5c34a66242d49746d8329dd1cdf27c965a9f17003a37ff584bc8293116d3"} Dec 12 07:03:00 crc kubenswrapper[4867]: I1212 07:03:00.834865 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-fbwh6" Dec 12 07:03:01 crc kubenswrapper[4867]: I1212 07:03:01.209631 4867 generic.go:334] "Generic (PLEG): container finished" podID="b966f81e-701c-49f5-bedf-732815bc91bb" containerID="aded5c34a66242d49746d8329dd1cdf27c965a9f17003a37ff584bc8293116d3" exitCode=0 Dec 12 07:03:01 crc kubenswrapper[4867]: I1212 07:03:01.209684 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerDied","Data":"aded5c34a66242d49746d8329dd1cdf27c965a9f17003a37ff584bc8293116d3"} Dec 12 07:03:01 crc kubenswrapper[4867]: I1212 07:03:01.250539 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-9wsvj" Dec 12 07:03:03 crc kubenswrapper[4867]: I1212 07:03:03.222464 4867 generic.go:334] "Generic (PLEG): container finished" podID="b966f81e-701c-49f5-bedf-732815bc91bb" containerID="dce4de9b9099b3d022355463d440f4ae7e41aaceb74a3700830f0d3b78094a5f" exitCode=0 Dec 12 07:03:03 crc kubenswrapper[4867]: I1212 07:03:03.222688 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerDied","Data":"dce4de9b9099b3d022355463d440f4ae7e41aaceb74a3700830f0d3b78094a5f"} Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.478799 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.560406 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util\") pod \"b966f81e-701c-49f5-bedf-732815bc91bb\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.560533 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zchh9\" (UniqueName: \"kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9\") pod \"b966f81e-701c-49f5-bedf-732815bc91bb\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.560602 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle\") pod \"b966f81e-701c-49f5-bedf-732815bc91bb\" (UID: \"b966f81e-701c-49f5-bedf-732815bc91bb\") " Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.561651 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle" (OuterVolumeSpecName: "bundle") pod "b966f81e-701c-49f5-bedf-732815bc91bb" (UID: "b966f81e-701c-49f5-bedf-732815bc91bb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.565786 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9" (OuterVolumeSpecName: "kube-api-access-zchh9") pod "b966f81e-701c-49f5-bedf-732815bc91bb" (UID: "b966f81e-701c-49f5-bedf-732815bc91bb"). InnerVolumeSpecName "kube-api-access-zchh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.570674 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util" (OuterVolumeSpecName: "util") pod "b966f81e-701c-49f5-bedf-732815bc91bb" (UID: "b966f81e-701c-49f5-bedf-732815bc91bb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.662335 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zchh9\" (UniqueName: \"kubernetes.io/projected/b966f81e-701c-49f5-bedf-732815bc91bb-kube-api-access-zchh9\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.662403 4867 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:04 crc kubenswrapper[4867]: I1212 07:03:04.662418 4867 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b966f81e-701c-49f5-bedf-732815bc91bb-util\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:05 crc kubenswrapper[4867]: I1212 07:03:05.236188 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" event={"ID":"b966f81e-701c-49f5-bedf-732815bc91bb","Type":"ContainerDied","Data":"917de5ad1718fe63719eb4494869c25416c107875ff8e9292d59e5eddfa41371"} Dec 12 07:03:05 crc kubenswrapper[4867]: I1212 07:03:05.236261 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="917de5ad1718fe63719eb4494869c25416c107875ff8e9292d59e5eddfa41371" Dec 12 07:03:05 crc kubenswrapper[4867]: I1212 07:03:05.236285 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.383721 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg"] Dec 12 07:03:08 crc kubenswrapper[4867]: E1212 07:03:08.384141 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="extract" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.384154 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="extract" Dec 12 07:03:08 crc kubenswrapper[4867]: E1212 07:03:08.384166 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="util" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.384171 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="util" Dec 12 07:03:08 crc kubenswrapper[4867]: E1212 07:03:08.384187 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="pull" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.384193 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="pull" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.384311 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b966f81e-701c-49f5-bedf-732815bc91bb" containerName="extract" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.384694 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.387045 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.387045 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.387494 4867 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-v487r" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.395482 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg"] Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.514097 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjqk6\" (UniqueName: \"kubernetes.io/projected/6480db6a-078e-444a-8e86-7c2ef0561fff-kube-api-access-jjqk6\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.514150 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6480db6a-078e-444a-8e86-7c2ef0561fff-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.615450 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjqk6\" (UniqueName: \"kubernetes.io/projected/6480db6a-078e-444a-8e86-7c2ef0561fff-kube-api-access-jjqk6\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.615522 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6480db6a-078e-444a-8e86-7c2ef0561fff-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.616086 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/6480db6a-078e-444a-8e86-7c2ef0561fff-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.641360 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjqk6\" (UniqueName: \"kubernetes.io/projected/6480db6a-078e-444a-8e86-7c2ef0561fff-kube-api-access-jjqk6\") pod \"cert-manager-operator-controller-manager-64cf6dff88-ttqlg\" (UID: \"6480db6a-078e-444a-8e86-7c2ef0561fff\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:08 crc kubenswrapper[4867]: I1212 07:03:08.705556 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" Dec 12 07:03:09 crc kubenswrapper[4867]: I1212 07:03:09.190999 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg"] Dec 12 07:03:09 crc kubenswrapper[4867]: W1212 07:03:09.195390 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6480db6a_078e_444a_8e86_7c2ef0561fff.slice/crio-c8f63b73a788b3b688f8e37365037a62efb54d9a582b02750cde085b5090a20b WatchSource:0}: Error finding container c8f63b73a788b3b688f8e37365037a62efb54d9a582b02750cde085b5090a20b: Status 404 returned error can't find the container with id c8f63b73a788b3b688f8e37365037a62efb54d9a582b02750cde085b5090a20b Dec 12 07:03:09 crc kubenswrapper[4867]: I1212 07:03:09.260833 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" event={"ID":"6480db6a-078e-444a-8e86-7c2ef0561fff","Type":"ContainerStarted","Data":"c8f63b73a788b3b688f8e37365037a62efb54d9a582b02750cde085b5090a20b"} Dec 12 07:03:10 crc kubenswrapper[4867]: I1212 07:03:10.539590 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-4rx5r" Dec 12 07:03:17 crc kubenswrapper[4867]: I1212 07:03:17.305389 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" event={"ID":"6480db6a-078e-444a-8e86-7c2ef0561fff","Type":"ContainerStarted","Data":"b4232df54f281003f198961ba45b35e08a4818dc8ecff1b738e897f461b55805"} Dec 12 07:03:17 crc kubenswrapper[4867]: I1212 07:03:17.323181 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-ttqlg" podStartSLOduration=2.227908553 podStartE2EDuration="9.323163027s" podCreationTimestamp="2025-12-12 07:03:08 +0000 UTC" firstStartedPulling="2025-12-12 07:03:09.198513142 +0000 UTC m=+876.769894411" lastFinishedPulling="2025-12-12 07:03:16.293767616 +0000 UTC m=+883.865148885" observedRunningTime="2025-12-12 07:03:17.320562672 +0000 UTC m=+884.891943961" watchObservedRunningTime="2025-12-12 07:03:17.323163027 +0000 UTC m=+884.894544306" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.619789 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-l4pl6"] Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.620875 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.623589 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.624465 4867 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6mnc9" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.624778 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.631193 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-l4pl6"] Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.766373 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.766447 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx2xr\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-kube-api-access-sx2xr\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.868417 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.868503 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx2xr\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-kube-api-access-sx2xr\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.887535 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx2xr\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-kube-api-access-sx2xr\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.890850 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/11283469-26a6-4a73-a330-e3427e0b22b1-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-l4pl6\" (UID: \"11283469-26a6-4a73-a330-e3427e0b22b1\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:19 crc kubenswrapper[4867]: I1212 07:03:19.937898 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:20 crc kubenswrapper[4867]: I1212 07:03:20.174619 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-l4pl6"] Dec 12 07:03:20 crc kubenswrapper[4867]: I1212 07:03:20.323821 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" event={"ID":"11283469-26a6-4a73-a330-e3427e0b22b1","Type":"ContainerStarted","Data":"4b8778732fcad1112ae8df1e56af1650395370057f43d78da8b1dee9db6b24ac"} Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.686979 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f"] Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.687962 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.728191 4867 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cfsrx" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.733622 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f"] Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.797528 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxcsv\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-kube-api-access-fxcsv\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.797663 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.899093 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxcsv\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-kube-api-access-fxcsv\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.899242 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.925020 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:21 crc kubenswrapper[4867]: I1212 07:03:21.942686 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxcsv\" (UniqueName: \"kubernetes.io/projected/492804a0-56c4-44eb-b9a6-1328cd8b7e84-kube-api-access-fxcsv\") pod \"cert-manager-cainjector-855d9ccff4-7ll2f\" (UID: \"492804a0-56c4-44eb-b9a6-1328cd8b7e84\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:22 crc kubenswrapper[4867]: I1212 07:03:22.047567 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" Dec 12 07:03:22 crc kubenswrapper[4867]: I1212 07:03:22.510624 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f"] Dec 12 07:03:23 crc kubenswrapper[4867]: I1212 07:03:23.363597 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" event={"ID":"492804a0-56c4-44eb-b9a6-1328cd8b7e84","Type":"ContainerStarted","Data":"2792833a797fec487653f0fa24ae7180e9f8b8a1fb0e4e32c1763967f01bdda3"} Dec 12 07:03:28 crc kubenswrapper[4867]: I1212 07:03:28.394024 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" event={"ID":"492804a0-56c4-44eb-b9a6-1328cd8b7e84","Type":"ContainerStarted","Data":"7e6d39b63722f3405638ac5395dcb467d83c9cd2f545f8869895e7dff64658fd"} Dec 12 07:03:28 crc kubenswrapper[4867]: I1212 07:03:28.396289 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" event={"ID":"11283469-26a6-4a73-a330-e3427e0b22b1","Type":"ContainerStarted","Data":"4388ba7055ef40d97bda2a67b066c2730dc670a3159e64f1b94b0eb3dfc69954"} Dec 12 07:03:28 crc kubenswrapper[4867]: I1212 07:03:28.396419 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:28 crc kubenswrapper[4867]: I1212 07:03:28.407913 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-7ll2f" podStartSLOduration=1.737407115 podStartE2EDuration="7.407891847s" podCreationTimestamp="2025-12-12 07:03:21 +0000 UTC" firstStartedPulling="2025-12-12 07:03:22.530790013 +0000 UTC m=+890.102171292" lastFinishedPulling="2025-12-12 07:03:28.201274755 +0000 UTC m=+895.772656024" observedRunningTime="2025-12-12 07:03:28.405451496 +0000 UTC m=+895.976832765" watchObservedRunningTime="2025-12-12 07:03:28.407891847 +0000 UTC m=+895.979273136" Dec 12 07:03:28 crc kubenswrapper[4867]: I1212 07:03:28.420941 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" podStartSLOduration=1.416368268 podStartE2EDuration="9.42091762s" podCreationTimestamp="2025-12-12 07:03:19 +0000 UTC" firstStartedPulling="2025-12-12 07:03:20.181628997 +0000 UTC m=+887.753010266" lastFinishedPulling="2025-12-12 07:03:28.186178349 +0000 UTC m=+895.757559618" observedRunningTime="2025-12-12 07:03:28.420584332 +0000 UTC m=+895.991965611" watchObservedRunningTime="2025-12-12 07:03:28.42091762 +0000 UTC m=+895.992298899" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.655746 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.657605 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.667159 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.777500 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.777566 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.777619 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5fd\" (UniqueName: \"kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.880509 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.880566 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.880603 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5fd\" (UniqueName: \"kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.881285 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.881670 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.900283 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5fd\" (UniqueName: \"kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd\") pod \"redhat-marketplace-2zkbq\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:31 crc kubenswrapper[4867]: I1212 07:03:31.979555 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:32 crc kubenswrapper[4867]: I1212 07:03:32.424423 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:33 crc kubenswrapper[4867]: I1212 07:03:33.425971 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerStarted","Data":"60bac031299aa4e39ec99bb595f320dc77ffe3fff71db93893bed23ea5b0318c"} Dec 12 07:03:34 crc kubenswrapper[4867]: I1212 07:03:34.433410 4867 generic.go:334] "Generic (PLEG): container finished" podID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerID="5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4" exitCode=0 Dec 12 07:03:34 crc kubenswrapper[4867]: I1212 07:03:34.433475 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerDied","Data":"5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4"} Dec 12 07:03:34 crc kubenswrapper[4867]: I1212 07:03:34.940131 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-l4pl6" Dec 12 07:03:36 crc kubenswrapper[4867]: I1212 07:03:36.460546 4867 generic.go:334] "Generic (PLEG): container finished" podID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerID="d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981" exitCode=0 Dec 12 07:03:36 crc kubenswrapper[4867]: I1212 07:03:36.460627 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerDied","Data":"d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981"} Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.503048 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-jm9f4"] Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.504083 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.507154 4867 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jz5gm" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.522052 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-jm9f4"] Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.668886 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-bound-sa-token\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.669406 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqlvg\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-kube-api-access-lqlvg\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.771338 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-bound-sa-token\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.771880 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqlvg\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-kube-api-access-lqlvg\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.793207 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqlvg\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-kube-api-access-lqlvg\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.794116 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3c8d2ab5-ca42-4bad-96fd-efa1b292cf70-bound-sa-token\") pod \"cert-manager-86cb77c54b-jm9f4\" (UID: \"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70\") " pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:38 crc kubenswrapper[4867]: I1212 07:03:38.820433 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-jm9f4" Dec 12 07:03:39 crc kubenswrapper[4867]: I1212 07:03:39.242795 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-jm9f4"] Dec 12 07:03:39 crc kubenswrapper[4867]: I1212 07:03:39.476907 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-jm9f4" event={"ID":"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70","Type":"ContainerStarted","Data":"3b96be49d7c305680dbc4fe36b0a21c8b41d8489dcc6e0f7867095465e8ea7e1"} Dec 12 07:03:43 crc kubenswrapper[4867]: I1212 07:03:43.506917 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerStarted","Data":"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7"} Dec 12 07:03:43 crc kubenswrapper[4867]: I1212 07:03:43.508559 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-jm9f4" event={"ID":"3c8d2ab5-ca42-4bad-96fd-efa1b292cf70","Type":"ContainerStarted","Data":"c2b71b849394be86db79f43e513abd74b046474ca33a042c8822b82f2d7d7930"} Dec 12 07:03:43 crc kubenswrapper[4867]: I1212 07:03:43.522320 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2zkbq" podStartSLOduration=4.112534285 podStartE2EDuration="12.522304192s" podCreationTimestamp="2025-12-12 07:03:31 +0000 UTC" firstStartedPulling="2025-12-12 07:03:34.436375791 +0000 UTC m=+902.007757060" lastFinishedPulling="2025-12-12 07:03:42.846145698 +0000 UTC m=+910.417526967" observedRunningTime="2025-12-12 07:03:43.520499919 +0000 UTC m=+911.091881188" watchObservedRunningTime="2025-12-12 07:03:43.522304192 +0000 UTC m=+911.093685451" Dec 12 07:03:43 crc kubenswrapper[4867]: I1212 07:03:43.542894 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-jm9f4" podStartSLOduration=5.5428737980000005 podStartE2EDuration="5.542873798s" podCreationTimestamp="2025-12-12 07:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:03:43.539111819 +0000 UTC m=+911.110493088" watchObservedRunningTime="2025-12-12 07:03:43.542873798 +0000 UTC m=+911.114255067" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.073337 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qz9fj"] Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.074702 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.077392 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.077632 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.077731 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-tnfkm" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.108826 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qz9fj"] Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.211129 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j78qp\" (UniqueName: \"kubernetes.io/projected/c06744d8-3f13-4940-9878-f2bf5a1eb5ee-kube-api-access-j78qp\") pod \"openstack-operator-index-qz9fj\" (UID: \"c06744d8-3f13-4940-9878-f2bf5a1eb5ee\") " pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.312290 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j78qp\" (UniqueName: \"kubernetes.io/projected/c06744d8-3f13-4940-9878-f2bf5a1eb5ee-kube-api-access-j78qp\") pod \"openstack-operator-index-qz9fj\" (UID: \"c06744d8-3f13-4940-9878-f2bf5a1eb5ee\") " pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.330338 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j78qp\" (UniqueName: \"kubernetes.io/projected/c06744d8-3f13-4940-9878-f2bf5a1eb5ee-kube-api-access-j78qp\") pod \"openstack-operator-index-qz9fj\" (UID: \"c06744d8-3f13-4940-9878-f2bf5a1eb5ee\") " pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.393556 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:48 crc kubenswrapper[4867]: I1212 07:03:48.784325 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qz9fj"] Dec 12 07:03:49 crc kubenswrapper[4867]: I1212 07:03:49.554812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qz9fj" event={"ID":"c06744d8-3f13-4940-9878-f2bf5a1eb5ee","Type":"ContainerStarted","Data":"6397546a148abfd826a72a7f2fa53cd0d4ee44a95c3a67482663cca2f9dee11f"} Dec 12 07:03:50 crc kubenswrapper[4867]: I1212 07:03:50.562165 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qz9fj" event={"ID":"c06744d8-3f13-4940-9878-f2bf5a1eb5ee","Type":"ContainerStarted","Data":"03bb7bfd8321913207cc4bdcfd64dfb12f5515d3eaa94d4d0a6f39164315fcb8"} Dec 12 07:03:50 crc kubenswrapper[4867]: I1212 07:03:50.575668 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qz9fj" podStartSLOduration=1.380480999 podStartE2EDuration="2.575649682s" podCreationTimestamp="2025-12-12 07:03:48 +0000 UTC" firstStartedPulling="2025-12-12 07:03:48.795813248 +0000 UTC m=+916.367194507" lastFinishedPulling="2025-12-12 07:03:49.990981931 +0000 UTC m=+917.562363190" observedRunningTime="2025-12-12 07:03:50.574538681 +0000 UTC m=+918.145919950" watchObservedRunningTime="2025-12-12 07:03:50.575649682 +0000 UTC m=+918.147030951" Dec 12 07:03:51 crc kubenswrapper[4867]: I1212 07:03:51.980421 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:51 crc kubenswrapper[4867]: I1212 07:03:51.981493 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:52 crc kubenswrapper[4867]: I1212 07:03:52.026727 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:52 crc kubenswrapper[4867]: I1212 07:03:52.610783 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.245244 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.247156 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.251566 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.392104 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.392205 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.392621 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mclx7\" (UniqueName: \"kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.494395 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mclx7\" (UniqueName: \"kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.494504 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.494536 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.495052 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.495072 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.515637 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mclx7\" (UniqueName: \"kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7\") pod \"certified-operators-9r8dn\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:55 crc kubenswrapper[4867]: I1212 07:03:55.605939 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.031791 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.035738 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.035955 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2zkbq" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="registry-server" containerID="cri-o://0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7" gracePeriod=2 Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.366670 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.405827 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content\") pod \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.405903 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj5fd\" (UniqueName: \"kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd\") pod \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.405951 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities\") pod \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\" (UID: \"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec\") " Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.406895 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities" (OuterVolumeSpecName: "utilities") pod "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" (UID: "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.411256 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd" (OuterVolumeSpecName: "kube-api-access-kj5fd") pod "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" (UID: "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec"). InnerVolumeSpecName "kube-api-access-kj5fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.424737 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" (UID: "34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.507146 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.507194 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj5fd\" (UniqueName: \"kubernetes.io/projected/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-kube-api-access-kj5fd\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.507208 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.597062 4867 generic.go:334] "Generic (PLEG): container finished" podID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerID="0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7" exitCode=0 Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.597133 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zkbq" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.597158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerDied","Data":"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7"} Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.597195 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zkbq" event={"ID":"34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec","Type":"ContainerDied","Data":"60bac031299aa4e39ec99bb595f320dc77ffe3fff71db93893bed23ea5b0318c"} Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.597244 4867 scope.go:117] "RemoveContainer" containerID="0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.599326 4867 generic.go:334] "Generic (PLEG): container finished" podID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerID="ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805" exitCode=0 Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.599360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerDied","Data":"ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805"} Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.599381 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerStarted","Data":"e178f95f628c3f187c955a3aa89f1843fcbcdfa753efe605435d782c33c48ac7"} Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.613084 4867 scope.go:117] "RemoveContainer" containerID="d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.628090 4867 scope.go:117] "RemoveContainer" containerID="5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.654401 4867 scope.go:117] "RemoveContainer" containerID="0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7" Dec 12 07:03:56 crc kubenswrapper[4867]: E1212 07:03:56.654933 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7\": container with ID starting with 0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7 not found: ID does not exist" containerID="0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.654989 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7"} err="failed to get container status \"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7\": rpc error: code = NotFound desc = could not find container \"0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7\": container with ID starting with 0c0a10aba91f1f361bf9635aef908c1e08c8b3100841d343d3e2268371395ee7 not found: ID does not exist" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.655017 4867 scope.go:117] "RemoveContainer" containerID="d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981" Dec 12 07:03:56 crc kubenswrapper[4867]: E1212 07:03:56.655508 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981\": container with ID starting with d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981 not found: ID does not exist" containerID="d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.655561 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981"} err="failed to get container status \"d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981\": rpc error: code = NotFound desc = could not find container \"d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981\": container with ID starting with d017a8a0dd0484c5e3e866b7eda8cda811c54cb402aec8fab2d3736504e22981 not found: ID does not exist" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.655597 4867 scope.go:117] "RemoveContainer" containerID="5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4" Dec 12 07:03:56 crc kubenswrapper[4867]: E1212 07:03:56.656004 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4\": container with ID starting with 5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4 not found: ID does not exist" containerID="5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.656050 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4"} err="failed to get container status \"5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4\": rpc error: code = NotFound desc = could not find container \"5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4\": container with ID starting with 5335a37f1612e5b6582e2400f5edeb06acd969b82274d5e471c2dbb1819d38a4 not found: ID does not exist" Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.682590 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.690729 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zkbq"] Dec 12 07:03:56 crc kubenswrapper[4867]: I1212 07:03:56.851140 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" path="/var/lib/kubelet/pods/34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec/volumes" Dec 12 07:03:57 crc kubenswrapper[4867]: I1212 07:03:57.607289 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerStarted","Data":"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88"} Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.394117 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.394172 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.420053 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.619914 4867 generic.go:334] "Generic (PLEG): container finished" podID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerID="f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88" exitCode=0 Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.619997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerDied","Data":"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88"} Dec 12 07:03:58 crc kubenswrapper[4867]: I1212 07:03:58.656032 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-qz9fj" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.888572 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw"] Dec 12 07:03:59 crc kubenswrapper[4867]: E1212 07:03:59.889064 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="extract-content" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.889076 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="extract-content" Dec 12 07:03:59 crc kubenswrapper[4867]: E1212 07:03:59.889087 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="extract-utilities" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.889093 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="extract-utilities" Dec 12 07:03:59 crc kubenswrapper[4867]: E1212 07:03:59.889106 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="registry-server" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.889112 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="registry-server" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.889244 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c489a6-f6f2-42bc-8eb5-e242c4a1a6ec" containerName="registry-server" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.890040 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.893682 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-ck5th" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.899077 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw"] Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.955772 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.955838 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:03:59 crc kubenswrapper[4867]: I1212 07:03:59.955935 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdjwb\" (UniqueName: \"kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.056748 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdjwb\" (UniqueName: \"kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.056852 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.056900 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.057439 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.057472 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.085647 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdjwb\" (UniqueName: \"kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb\") pod \"65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.203527 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.634175 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerStarted","Data":"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f"} Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.663508 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9r8dn" podStartSLOduration=2.657552533 podStartE2EDuration="5.663489257s" podCreationTimestamp="2025-12-12 07:03:55 +0000 UTC" firstStartedPulling="2025-12-12 07:03:56.600923013 +0000 UTC m=+924.172304282" lastFinishedPulling="2025-12-12 07:03:59.606859737 +0000 UTC m=+927.178241006" observedRunningTime="2025-12-12 07:04:00.655786466 +0000 UTC m=+928.227167735" watchObservedRunningTime="2025-12-12 07:04:00.663489257 +0000 UTC m=+928.234870526" Dec 12 07:04:00 crc kubenswrapper[4867]: I1212 07:04:00.666306 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw"] Dec 12 07:04:01 crc kubenswrapper[4867]: I1212 07:04:01.640705 4867 generic.go:334] "Generic (PLEG): container finished" podID="a40a7208-8353-437c-900b-6341f482a91a" containerID="3049296399bfbb48b3eddd52b2ab6f318f358c64ab46bb3d1f4e422624a3731d" exitCode=0 Dec 12 07:04:01 crc kubenswrapper[4867]: I1212 07:04:01.640827 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" event={"ID":"a40a7208-8353-437c-900b-6341f482a91a","Type":"ContainerDied","Data":"3049296399bfbb48b3eddd52b2ab6f318f358c64ab46bb3d1f4e422624a3731d"} Dec 12 07:04:01 crc kubenswrapper[4867]: I1212 07:04:01.642318 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" event={"ID":"a40a7208-8353-437c-900b-6341f482a91a","Type":"ContainerStarted","Data":"6d3728ac7e2da256c6bb71c4386bc4a26595eab8e640f8f8cec7277b9c54cc94"} Dec 12 07:04:03 crc kubenswrapper[4867]: I1212 07:04:03.657852 4867 generic.go:334] "Generic (PLEG): container finished" podID="a40a7208-8353-437c-900b-6341f482a91a" containerID="1008854b7ef7ebb46c582648c23c528da66e1db05c67f658af5588d18e76fee2" exitCode=0 Dec 12 07:04:03 crc kubenswrapper[4867]: I1212 07:04:03.658087 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" event={"ID":"a40a7208-8353-437c-900b-6341f482a91a","Type":"ContainerDied","Data":"1008854b7ef7ebb46c582648c23c528da66e1db05c67f658af5588d18e76fee2"} Dec 12 07:04:04 crc kubenswrapper[4867]: I1212 07:04:04.669738 4867 generic.go:334] "Generic (PLEG): container finished" podID="a40a7208-8353-437c-900b-6341f482a91a" containerID="39c5338186fa6fc92ec35083662a7731d33c73c39fccde7c0ffb4c262e2d8ccc" exitCode=0 Dec 12 07:04:04 crc kubenswrapper[4867]: I1212 07:04:04.669787 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" event={"ID":"a40a7208-8353-437c-900b-6341f482a91a","Type":"ContainerDied","Data":"39c5338186fa6fc92ec35083662a7731d33c73c39fccde7c0ffb4c262e2d8ccc"} Dec 12 07:04:05 crc kubenswrapper[4867]: I1212 07:04:05.606781 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:05 crc kubenswrapper[4867]: I1212 07:04:05.607080 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:05 crc kubenswrapper[4867]: I1212 07:04:05.647334 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:05 crc kubenswrapper[4867]: I1212 07:04:05.715431 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:05 crc kubenswrapper[4867]: I1212 07:04:05.914603 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.036154 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle\") pod \"a40a7208-8353-437c-900b-6341f482a91a\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.036230 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdjwb\" (UniqueName: \"kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb\") pod \"a40a7208-8353-437c-900b-6341f482a91a\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.036377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util\") pod \"a40a7208-8353-437c-900b-6341f482a91a\" (UID: \"a40a7208-8353-437c-900b-6341f482a91a\") " Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.036934 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle" (OuterVolumeSpecName: "bundle") pod "a40a7208-8353-437c-900b-6341f482a91a" (UID: "a40a7208-8353-437c-900b-6341f482a91a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.042420 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb" (OuterVolumeSpecName: "kube-api-access-qdjwb") pod "a40a7208-8353-437c-900b-6341f482a91a" (UID: "a40a7208-8353-437c-900b-6341f482a91a"). InnerVolumeSpecName "kube-api-access-qdjwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.050027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util" (OuterVolumeSpecName: "util") pod "a40a7208-8353-437c-900b-6341f482a91a" (UID: "a40a7208-8353-437c-900b-6341f482a91a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.137677 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdjwb\" (UniqueName: \"kubernetes.io/projected/a40a7208-8353-437c-900b-6341f482a91a-kube-api-access-qdjwb\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.137921 4867 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-util\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.137982 4867 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a40a7208-8353-437c-900b-6341f482a91a-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.687483 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.688223 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw" event={"ID":"a40a7208-8353-437c-900b-6341f482a91a","Type":"ContainerDied","Data":"6d3728ac7e2da256c6bb71c4386bc4a26595eab8e640f8f8cec7277b9c54cc94"} Dec 12 07:04:06 crc kubenswrapper[4867]: I1212 07:04:06.688366 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d3728ac7e2da256c6bb71c4386bc4a26595eab8e640f8f8cec7277b9c54cc94" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.233107 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.642047 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:07 crc kubenswrapper[4867]: E1212 07:04:07.642490 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="util" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.642582 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="util" Dec 12 07:04:07 crc kubenswrapper[4867]: E1212 07:04:07.642649 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="extract" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.642713 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="extract" Dec 12 07:04:07 crc kubenswrapper[4867]: E1212 07:04:07.642771 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="pull" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.642825 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="pull" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.643002 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40a7208-8353-437c-900b-6341f482a91a" containerName="extract" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.643850 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.665497 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.758557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.758770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.758805 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b8qx\" (UniqueName: \"kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.859647 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.859694 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b8qx\" (UniqueName: \"kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.859755 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.860489 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.860577 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.878368 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b8qx\" (UniqueName: \"kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx\") pod \"community-operators-sgwz9\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:07 crc kubenswrapper[4867]: I1212 07:04:07.957876 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:08 crc kubenswrapper[4867]: I1212 07:04:08.352578 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:08 crc kubenswrapper[4867]: W1212 07:04:08.356688 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63167195_e501_454c_a5dc_9d5c21f4b26a.slice/crio-3d1bb17676acdde86442fec5886e6c818a9b6d46dc9d47745020b5ad62681859 WatchSource:0}: Error finding container 3d1bb17676acdde86442fec5886e6c818a9b6d46dc9d47745020b5ad62681859: Status 404 returned error can't find the container with id 3d1bb17676acdde86442fec5886e6c818a9b6d46dc9d47745020b5ad62681859 Dec 12 07:04:08 crc kubenswrapper[4867]: I1212 07:04:08.701932 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerStarted","Data":"3d1bb17676acdde86442fec5886e6c818a9b6d46dc9d47745020b5ad62681859"} Dec 12 07:04:08 crc kubenswrapper[4867]: I1212 07:04:08.702111 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9r8dn" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="registry-server" containerID="cri-o://0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f" gracePeriod=2 Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.559772 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.578671 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content\") pod \"191452f7-1122-4fb2-b467-b5f0ad27f274\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.627057 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "191452f7-1122-4fb2-b467-b5f0ad27f274" (UID: "191452f7-1122-4fb2-b467-b5f0ad27f274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.679571 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mclx7\" (UniqueName: \"kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7\") pod \"191452f7-1122-4fb2-b467-b5f0ad27f274\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.679718 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities\") pod \"191452f7-1122-4fb2-b467-b5f0ad27f274\" (UID: \"191452f7-1122-4fb2-b467-b5f0ad27f274\") " Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.680145 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.680360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities" (OuterVolumeSpecName: "utilities") pod "191452f7-1122-4fb2-b467-b5f0ad27f274" (UID: "191452f7-1122-4fb2-b467-b5f0ad27f274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.684327 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7" (OuterVolumeSpecName: "kube-api-access-mclx7") pod "191452f7-1122-4fb2-b467-b5f0ad27f274" (UID: "191452f7-1122-4fb2-b467-b5f0ad27f274"). InnerVolumeSpecName "kube-api-access-mclx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.710161 4867 generic.go:334] "Generic (PLEG): container finished" podID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerID="0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f" exitCode=0 Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.710219 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9r8dn" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.710302 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerDied","Data":"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f"} Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.710370 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9r8dn" event={"ID":"191452f7-1122-4fb2-b467-b5f0ad27f274","Type":"ContainerDied","Data":"e178f95f628c3f187c955a3aa89f1843fcbcdfa753efe605435d782c33c48ac7"} Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.710393 4867 scope.go:117] "RemoveContainer" containerID="0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.713513 4867 generic.go:334] "Generic (PLEG): container finished" podID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerID="1ebf7c75b60d56473ec05213fd69a9bff135ec6fd498647a831db2481489e760" exitCode=0 Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.713550 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerDied","Data":"1ebf7c75b60d56473ec05213fd69a9bff135ec6fd498647a831db2481489e760"} Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.736277 4867 scope.go:117] "RemoveContainer" containerID="f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.753365 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.757293 4867 scope.go:117] "RemoveContainer" containerID="ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.759097 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9r8dn"] Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.775615 4867 scope.go:117] "RemoveContainer" containerID="0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f" Dec 12 07:04:09 crc kubenswrapper[4867]: E1212 07:04:09.776417 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f\": container with ID starting with 0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f not found: ID does not exist" containerID="0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.776512 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f"} err="failed to get container status \"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f\": rpc error: code = NotFound desc = could not find container \"0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f\": container with ID starting with 0044587741f862e7006f129fa1a976b9e269a5dbae65110e3bd677f17b2efe3f not found: ID does not exist" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.776549 4867 scope.go:117] "RemoveContainer" containerID="f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88" Dec 12 07:04:09 crc kubenswrapper[4867]: E1212 07:04:09.777021 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88\": container with ID starting with f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88 not found: ID does not exist" containerID="f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.777063 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88"} err="failed to get container status \"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88\": rpc error: code = NotFound desc = could not find container \"f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88\": container with ID starting with f3861240def7f0687c89d0de0fc95371fd262791643c3e953cec2b6740c84d88 not found: ID does not exist" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.777095 4867 scope.go:117] "RemoveContainer" containerID="ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805" Dec 12 07:04:09 crc kubenswrapper[4867]: E1212 07:04:09.777596 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805\": container with ID starting with ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805 not found: ID does not exist" containerID="ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.777647 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805"} err="failed to get container status \"ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805\": rpc error: code = NotFound desc = could not find container \"ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805\": container with ID starting with ec3020528f06546ebee5d35d96655a8d6cfd117915479ef6631195d678040805 not found: ID does not exist" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.782006 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mclx7\" (UniqueName: \"kubernetes.io/projected/191452f7-1122-4fb2-b467-b5f0ad27f274-kube-api-access-mclx7\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:09 crc kubenswrapper[4867]: I1212 07:04:09.782032 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/191452f7-1122-4fb2-b467-b5f0ad27f274-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:10 crc kubenswrapper[4867]: I1212 07:04:10.721668 4867 generic.go:334] "Generic (PLEG): container finished" podID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerID="e669966bd2e3727324e8038e5e3522ff0d04ff16d5bdaaeb46e0498f6a663a03" exitCode=0 Dec 12 07:04:10 crc kubenswrapper[4867]: I1212 07:04:10.721741 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerDied","Data":"e669966bd2e3727324e8038e5e3522ff0d04ff16d5bdaaeb46e0498f6a663a03"} Dec 12 07:04:10 crc kubenswrapper[4867]: I1212 07:04:10.845593 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" path="/var/lib/kubelet/pods/191452f7-1122-4fb2-b467-b5f0ad27f274/volumes" Dec 12 07:04:11 crc kubenswrapper[4867]: I1212 07:04:11.729790 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerStarted","Data":"e565780f1d07419dfe71188083b5ab987ed4976147e860f231421af85d559d28"} Dec 12 07:04:11 crc kubenswrapper[4867]: I1212 07:04:11.748170 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sgwz9" podStartSLOduration=2.9508017779999998 podStartE2EDuration="4.748146363s" podCreationTimestamp="2025-12-12 07:04:07 +0000 UTC" firstStartedPulling="2025-12-12 07:04:09.715000986 +0000 UTC m=+937.286382255" lastFinishedPulling="2025-12-12 07:04:11.512345571 +0000 UTC m=+939.083726840" observedRunningTime="2025-12-12 07:04:11.744616998 +0000 UTC m=+939.315998287" watchObservedRunningTime="2025-12-12 07:04:11.748146363 +0000 UTC m=+939.319527642" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.107181 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt"] Dec 12 07:04:12 crc kubenswrapper[4867]: E1212 07:04:12.107974 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="registry-server" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.107990 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="registry-server" Dec 12 07:04:12 crc kubenswrapper[4867]: E1212 07:04:12.108022 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="extract-utilities" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.108032 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="extract-utilities" Dec 12 07:04:12 crc kubenswrapper[4867]: E1212 07:04:12.108053 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="extract-content" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.108063 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="extract-content" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.108419 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="191452f7-1122-4fb2-b467-b5f0ad27f274" containerName="registry-server" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.109083 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.110873 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-62qdr" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.148378 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt"] Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.215415 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc2jc\" (UniqueName: \"kubernetes.io/projected/98b6024c-e03c-45a1-856b-e3f4af5483cd-kube-api-access-zc2jc\") pod \"openstack-operator-controller-operator-799b66f579-q8vtt\" (UID: \"98b6024c-e03c-45a1-856b-e3f4af5483cd\") " pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.317181 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc2jc\" (UniqueName: \"kubernetes.io/projected/98b6024c-e03c-45a1-856b-e3f4af5483cd-kube-api-access-zc2jc\") pod \"openstack-operator-controller-operator-799b66f579-q8vtt\" (UID: \"98b6024c-e03c-45a1-856b-e3f4af5483cd\") " pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.336700 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc2jc\" (UniqueName: \"kubernetes.io/projected/98b6024c-e03c-45a1-856b-e3f4af5483cd-kube-api-access-zc2jc\") pod \"openstack-operator-controller-operator-799b66f579-q8vtt\" (UID: \"98b6024c-e03c-45a1-856b-e3f4af5483cd\") " pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.430684 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.669679 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt"] Dec 12 07:04:12 crc kubenswrapper[4867]: I1212 07:04:12.736887 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" event={"ID":"98b6024c-e03c-45a1-856b-e3f4af5483cd","Type":"ContainerStarted","Data":"e5cd077f518e86e077f03bdac22916f9d01104ce85e57bd1e07ac485bf9308e5"} Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.776265 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" event={"ID":"98b6024c-e03c-45a1-856b-e3f4af5483cd","Type":"ContainerStarted","Data":"0347f633d9c387302b7c5e59f3e7f87a41f68c580622ab4454a22b1f41525be7"} Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.776899 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.810456 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" podStartSLOduration=1.002394026 podStartE2EDuration="5.810439571s" podCreationTimestamp="2025-12-12 07:04:12 +0000 UTC" firstStartedPulling="2025-12-12 07:04:12.674671284 +0000 UTC m=+940.246052553" lastFinishedPulling="2025-12-12 07:04:17.482716829 +0000 UTC m=+945.054098098" observedRunningTime="2025-12-12 07:04:17.805652764 +0000 UTC m=+945.377034043" watchObservedRunningTime="2025-12-12 07:04:17.810439571 +0000 UTC m=+945.381820840" Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.958720 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.958773 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:17 crc kubenswrapper[4867]: I1212 07:04:17.996877 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:18 crc kubenswrapper[4867]: I1212 07:04:18.824500 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:20 crc kubenswrapper[4867]: I1212 07:04:20.233394 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:20 crc kubenswrapper[4867]: I1212 07:04:20.793912 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sgwz9" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="registry-server" containerID="cri-o://e565780f1d07419dfe71188083b5ab987ed4976147e860f231421af85d559d28" gracePeriod=2 Dec 12 07:04:21 crc kubenswrapper[4867]: I1212 07:04:21.801614 4867 generic.go:334] "Generic (PLEG): container finished" podID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerID="e565780f1d07419dfe71188083b5ab987ed4976147e860f231421af85d559d28" exitCode=0 Dec 12 07:04:21 crc kubenswrapper[4867]: I1212 07:04:21.801663 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerDied","Data":"e565780f1d07419dfe71188083b5ab987ed4976147e860f231421af85d559d28"} Dec 12 07:04:22 crc kubenswrapper[4867]: I1212 07:04:22.433288 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-799b66f579-q8vtt" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.083495 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.164984 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b8qx\" (UniqueName: \"kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx\") pod \"63167195-e501-454c-a5dc-9d5c21f4b26a\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.165054 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content\") pod \"63167195-e501-454c-a5dc-9d5c21f4b26a\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.165179 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities\") pod \"63167195-e501-454c-a5dc-9d5c21f4b26a\" (UID: \"63167195-e501-454c-a5dc-9d5c21f4b26a\") " Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.166377 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities" (OuterVolumeSpecName: "utilities") pod "63167195-e501-454c-a5dc-9d5c21f4b26a" (UID: "63167195-e501-454c-a5dc-9d5c21f4b26a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.172460 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx" (OuterVolumeSpecName: "kube-api-access-6b8qx") pod "63167195-e501-454c-a5dc-9d5c21f4b26a" (UID: "63167195-e501-454c-a5dc-9d5c21f4b26a"). InnerVolumeSpecName "kube-api-access-6b8qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.232359 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63167195-e501-454c-a5dc-9d5c21f4b26a" (UID: "63167195-e501-454c-a5dc-9d5c21f4b26a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.266245 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b8qx\" (UniqueName: \"kubernetes.io/projected/63167195-e501-454c-a5dc-9d5c21f4b26a-kube-api-access-6b8qx\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.266280 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.266292 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63167195-e501-454c-a5dc-9d5c21f4b26a-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.817866 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgwz9" event={"ID":"63167195-e501-454c-a5dc-9d5c21f4b26a","Type":"ContainerDied","Data":"3d1bb17676acdde86442fec5886e6c818a9b6d46dc9d47745020b5ad62681859"} Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.818016 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgwz9" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.818204 4867 scope.go:117] "RemoveContainer" containerID="e565780f1d07419dfe71188083b5ab987ed4976147e860f231421af85d559d28" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.836048 4867 scope.go:117] "RemoveContainer" containerID="e669966bd2e3727324e8038e5e3522ff0d04ff16d5bdaaeb46e0498f6a663a03" Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.846452 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.851327 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sgwz9"] Dec 12 07:04:23 crc kubenswrapper[4867]: I1212 07:04:23.873041 4867 scope.go:117] "RemoveContainer" containerID="1ebf7c75b60d56473ec05213fd69a9bff135ec6fd498647a831db2481489e760" Dec 12 07:04:24 crc kubenswrapper[4867]: I1212 07:04:24.846189 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" path="/var/lib/kubelet/pods/63167195-e501-454c-a5dc-9d5c21f4b26a/volumes" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.946564 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd"] Dec 12 07:04:40 crc kubenswrapper[4867]: E1212 07:04:40.947525 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="extract-utilities" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.947544 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="extract-utilities" Dec 12 07:04:40 crc kubenswrapper[4867]: E1212 07:04:40.947571 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="extract-content" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.947579 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="extract-content" Dec 12 07:04:40 crc kubenswrapper[4867]: E1212 07:04:40.947594 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="registry-server" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.947602 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="registry-server" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.947751 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="63167195-e501-454c-a5dc-9d5c21f4b26a" containerName="registry-server" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.948617 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.954934 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd"] Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.954961 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-86kc4" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.962532 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk"] Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.968539 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.970075 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-b5lrk" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.980390 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2"] Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.981929 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.983708 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5cc9x" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.991286 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk"] Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.996835 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtvq\" (UniqueName: \"kubernetes.io/projected/d774c064-6d06-4d04-8c11-7008260c0044-kube-api-access-ldtvq\") pod \"barbican-operator-controller-manager-7d9dfd778-9rhfd\" (UID: \"d774c064-6d06-4d04-8c11-7008260c0044\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:04:40 crc kubenswrapper[4867]: I1212 07:04:40.997026 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w2w8\" (UniqueName: \"kubernetes.io/projected/2962c7e5-a1d8-451b-9d65-f4e6e8295e2f-kube-api-access-7w2w8\") pod \"designate-operator-controller-manager-697fb699cf-kf2v2\" (UID: \"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.007276 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.027089 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.028050 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.034817 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-n79lj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.049535 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.064173 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.065375 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.067671 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-fntsv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.072010 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.076834 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.078072 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.080988 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mgc4l" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.101118 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtvq\" (UniqueName: \"kubernetes.io/projected/d774c064-6d06-4d04-8c11-7008260c0044-kube-api-access-ldtvq\") pod \"barbican-operator-controller-manager-7d9dfd778-9rhfd\" (UID: \"d774c064-6d06-4d04-8c11-7008260c0044\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.101196 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.101390 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfbpc\" (UniqueName: \"kubernetes.io/projected/8731ddb0-38b0-4de8-895a-c880c8899d26-kube-api-access-pfbpc\") pod \"cinder-operator-controller-manager-6c677c69b-5zfsk\" (UID: \"8731ddb0-38b0-4de8-895a-c880c8899d26\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.101440 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w2w8\" (UniqueName: \"kubernetes.io/projected/2962c7e5-a1d8-451b-9d65-f4e6e8295e2f-kube-api-access-7w2w8\") pod \"designate-operator-controller-manager-697fb699cf-kf2v2\" (UID: \"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.102718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.106865 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.107962 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-gr8pz" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.108779 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.116447 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.117899 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.119976 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-5bnfw" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.133523 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w2w8\" (UniqueName: \"kubernetes.io/projected/2962c7e5-a1d8-451b-9d65-f4e6e8295e2f-kube-api-access-7w2w8\") pod \"designate-operator-controller-manager-697fb699cf-kf2v2\" (UID: \"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.133577 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtvq\" (UniqueName: \"kubernetes.io/projected/d774c064-6d06-4d04-8c11-7008260c0044-kube-api-access-ldtvq\") pod \"barbican-operator-controller-manager-7d9dfd778-9rhfd\" (UID: \"d774c064-6d06-4d04-8c11-7008260c0044\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.137315 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.140487 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.141434 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.143834 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6sdj8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.157299 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.163210 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.164093 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.167753 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.168525 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-xb5wp" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.177664 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.204615 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.213678 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfbpc\" (UniqueName: \"kubernetes.io/projected/8731ddb0-38b0-4de8-895a-c880c8899d26-kube-api-access-pfbpc\") pod \"cinder-operator-controller-manager-6c677c69b-5zfsk\" (UID: \"8731ddb0-38b0-4de8-895a-c880c8899d26\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.213758 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px72p\" (UniqueName: \"kubernetes.io/projected/93fd7786-fb3e-45aa-92d2-6b6fa38daf56-kube-api-access-px72p\") pod \"horizon-operator-controller-manager-68c6d99b8f-z9jcv\" (UID: \"93fd7786-fb3e-45aa-92d2-6b6fa38daf56\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.213795 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8x26\" (UniqueName: \"kubernetes.io/projected/09531d04-5bd7-4457-af43-296c2e76cdfb-kube-api-access-s8x26\") pod \"heat-operator-controller-manager-5f64f6f8bb-qwkr9\" (UID: \"09531d04-5bd7-4457-af43-296c2e76cdfb\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.213834 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj9p2\" (UniqueName: \"kubernetes.io/projected/528acf70-c2ad-4968-88ad-24531e8f5d12-kube-api-access-fj9p2\") pod \"glance-operator-controller-manager-5697bb5779-5mwnj\" (UID: \"528acf70-c2ad-4968-88ad-24531e8f5d12\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.216701 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.228562 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-hjcz6" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.276811 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.279319 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.283879 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfbpc\" (UniqueName: \"kubernetes.io/projected/8731ddb0-38b0-4de8-895a-c880c8899d26-kube-api-access-pfbpc\") pod \"cinder-operator-controller-manager-6c677c69b-5zfsk\" (UID: \"8731ddb0-38b0-4de8-895a-c880c8899d26\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.289731 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.306450 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.313178 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.314358 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj9p2\" (UniqueName: \"kubernetes.io/projected/528acf70-c2ad-4968-88ad-24531e8f5d12-kube-api-access-fj9p2\") pod \"glance-operator-controller-manager-5697bb5779-5mwnj\" (UID: \"528acf70-c2ad-4968-88ad-24531e8f5d12\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315413 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzhwh\" (UniqueName: \"kubernetes.io/projected/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-kube-api-access-rzhwh\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgxnr\" (UniqueName: \"kubernetes.io/projected/f70067b7-bc5e-4f40-9b90-7601ac37aa89-kube-api-access-dgxnr\") pod \"mariadb-operator-controller-manager-79c8c4686c-frb6j\" (UID: \"f70067b7-bc5e-4f40-9b90-7601ac37aa89\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315516 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px72p\" (UniqueName: \"kubernetes.io/projected/93fd7786-fb3e-45aa-92d2-6b6fa38daf56-kube-api-access-px72p\") pod \"horizon-operator-controller-manager-68c6d99b8f-z9jcv\" (UID: \"93fd7786-fb3e-45aa-92d2-6b6fa38daf56\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315542 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwbz\" (UniqueName: \"kubernetes.io/projected/eacc6270-8929-44a1-a971-f2a6b7582103-kube-api-access-6xwbz\") pod \"ironic-operator-controller-manager-967d97867-rfkwx\" (UID: \"eacc6270-8929-44a1-a971-f2a6b7582103\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315572 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsgzg\" (UniqueName: \"kubernetes.io/projected/9c8a7928-c0c9-4516-8e34-273afc40ccae-kube-api-access-dsgzg\") pod \"keystone-operator-controller-manager-7765d96ddf-62qkh\" (UID: \"9c8a7928-c0c9-4516-8e34-273afc40ccae\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8x26\" (UniqueName: \"kubernetes.io/projected/09531d04-5bd7-4457-af43-296c2e76cdfb-kube-api-access-s8x26\") pod \"heat-operator-controller-manager-5f64f6f8bb-qwkr9\" (UID: \"09531d04-5bd7-4457-af43-296c2e76cdfb\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315618 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7wbf\" (UniqueName: \"kubernetes.io/projected/eab0a735-13ed-4d91-a4c5-d3c09d99139d-kube-api-access-b7wbf\") pod \"manila-operator-controller-manager-5b5fd79c9c-2hqbr\" (UID: \"eab0a735-13ed-4d91-a4c5-d3c09d99139d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.315638 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.317984 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-v9wj6" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.337548 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.338825 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.356848 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-wbxgz" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.360892 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px72p\" (UniqueName: \"kubernetes.io/projected/93fd7786-fb3e-45aa-92d2-6b6fa38daf56-kube-api-access-px72p\") pod \"horizon-operator-controller-manager-68c6d99b8f-z9jcv\" (UID: \"93fd7786-fb3e-45aa-92d2-6b6fa38daf56\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.362121 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8x26\" (UniqueName: \"kubernetes.io/projected/09531d04-5bd7-4457-af43-296c2e76cdfb-kube-api-access-s8x26\") pod \"heat-operator-controller-manager-5f64f6f8bb-qwkr9\" (UID: \"09531d04-5bd7-4457-af43-296c2e76cdfb\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.370857 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj9p2\" (UniqueName: \"kubernetes.io/projected/528acf70-c2ad-4968-88ad-24531e8f5d12-kube-api-access-fj9p2\") pod \"glance-operator-controller-manager-5697bb5779-5mwnj\" (UID: \"528acf70-c2ad-4968-88ad-24531e8f5d12\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.387558 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.405582 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417317 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwbz\" (UniqueName: \"kubernetes.io/projected/eacc6270-8929-44a1-a971-f2a6b7582103-kube-api-access-6xwbz\") pod \"ironic-operator-controller-manager-967d97867-rfkwx\" (UID: \"eacc6270-8929-44a1-a971-f2a6b7582103\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417365 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgzg\" (UniqueName: \"kubernetes.io/projected/9c8a7928-c0c9-4516-8e34-273afc40ccae-kube-api-access-dsgzg\") pod \"keystone-operator-controller-manager-7765d96ddf-62qkh\" (UID: \"9c8a7928-c0c9-4516-8e34-273afc40ccae\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417394 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7wbf\" (UniqueName: \"kubernetes.io/projected/eab0a735-13ed-4d91-a4c5-d3c09d99139d-kube-api-access-b7wbf\") pod \"manila-operator-controller-manager-5b5fd79c9c-2hqbr\" (UID: \"eab0a735-13ed-4d91-a4c5-d3c09d99139d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417417 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417445 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dmwt\" (UniqueName: \"kubernetes.io/projected/09951d5e-0bdd-4a76-b433-880fa1bb7abd-kube-api-access-8dmwt\") pod \"nova-operator-controller-manager-697bc559fc-6drsk\" (UID: \"09951d5e-0bdd-4a76-b433-880fa1bb7abd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417462 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtkfp\" (UniqueName: \"kubernetes.io/projected/5a7bb06a-0518-4384-b694-b73194df4a05-kube-api-access-vtkfp\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-fhp5f\" (UID: \"5a7bb06a-0518-4384-b694-b73194df4a05\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417494 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzhwh\" (UniqueName: \"kubernetes.io/projected/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-kube-api-access-rzhwh\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.417514 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgxnr\" (UniqueName: \"kubernetes.io/projected/f70067b7-bc5e-4f40-9b90-7601ac37aa89-kube-api-access-dgxnr\") pod \"mariadb-operator-controller-manager-79c8c4686c-frb6j\" (UID: \"f70067b7-bc5e-4f40-9b90-7601ac37aa89\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.418168 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.418209 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:41.91819391 +0000 UTC m=+969.489575179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.421306 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.440311 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.454334 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgxnr\" (UniqueName: \"kubernetes.io/projected/f70067b7-bc5e-4f40-9b90-7601ac37aa89-kube-api-access-dgxnr\") pod \"mariadb-operator-controller-manager-79c8c4686c-frb6j\" (UID: \"f70067b7-bc5e-4f40-9b90-7601ac37aa89\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.460379 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.461877 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.464175 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzhwh\" (UniqueName: \"kubernetes.io/projected/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-kube-api-access-rzhwh\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.468525 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-4sqjr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.473679 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwbz\" (UniqueName: \"kubernetes.io/projected/eacc6270-8929-44a1-a971-f2a6b7582103-kube-api-access-6xwbz\") pod \"ironic-operator-controller-manager-967d97867-rfkwx\" (UID: \"eacc6270-8929-44a1-a971-f2a6b7582103\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.481749 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7wbf\" (UniqueName: \"kubernetes.io/projected/eab0a735-13ed-4d91-a4c5-d3c09d99139d-kube-api-access-b7wbf\") pod \"manila-operator-controller-manager-5b5fd79c9c-2hqbr\" (UID: \"eab0a735-13ed-4d91-a4c5-d3c09d99139d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.486327 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgzg\" (UniqueName: \"kubernetes.io/projected/9c8a7928-c0c9-4516-8e34-273afc40ccae-kube-api-access-dsgzg\") pod \"keystone-operator-controller-manager-7765d96ddf-62qkh\" (UID: \"9c8a7928-c0c9-4516-8e34-273afc40ccae\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.522135 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.532440 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dmwt\" (UniqueName: \"kubernetes.io/projected/09951d5e-0bdd-4a76-b433-880fa1bb7abd-kube-api-access-8dmwt\") pod \"nova-operator-controller-manager-697bc559fc-6drsk\" (UID: \"09951d5e-0bdd-4a76-b433-880fa1bb7abd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.532474 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtkfp\" (UniqueName: \"kubernetes.io/projected/5a7bb06a-0518-4384-b694-b73194df4a05-kube-api-access-vtkfp\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-fhp5f\" (UID: \"5a7bb06a-0518-4384-b694-b73194df4a05\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.532615 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgrh\" (UniqueName: \"kubernetes.io/projected/b29bfc3e-61ea-443a-af46-13583b7d619f-kube-api-access-whgrh\") pod \"octavia-operator-controller-manager-998648c74-c8dlk\" (UID: \"b29bfc3e-61ea-443a-af46-13583b7d619f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.564647 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dmwt\" (UniqueName: \"kubernetes.io/projected/09951d5e-0bdd-4a76-b433-880fa1bb7abd-kube-api-access-8dmwt\") pod \"nova-operator-controller-manager-697bc559fc-6drsk\" (UID: \"09951d5e-0bdd-4a76-b433-880fa1bb7abd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.578783 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtkfp\" (UniqueName: \"kubernetes.io/projected/5a7bb06a-0518-4384-b694-b73194df4a05-kube-api-access-vtkfp\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-fhp5f\" (UID: \"5a7bb06a-0518-4384-b694-b73194df4a05\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.582559 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.589552 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.633770 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.634823 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgrh\" (UniqueName: \"kubernetes.io/projected/b29bfc3e-61ea-443a-af46-13583b7d619f-kube-api-access-whgrh\") pod \"octavia-operator-controller-manager-998648c74-c8dlk\" (UID: \"b29bfc3e-61ea-443a-af46-13583b7d619f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.634864 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.640571 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sdvlm" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.649658 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.656381 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgrh\" (UniqueName: \"kubernetes.io/projected/b29bfc3e-61ea-443a-af46-13583b7d619f-kube-api-access-whgrh\") pod \"octavia-operator-controller-manager-998648c74-c8dlk\" (UID: \"b29bfc3e-61ea-443a-af46-13583b7d619f\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.663919 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.665097 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.667169 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hd9b8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.667352 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.670320 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.683380 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.712548 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.712930 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.714682 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.735848 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx9fg\" (UniqueName: \"kubernetes.io/projected/42800642-ed16-4b36-8587-8d136d065283-kube-api-access-zx9fg\") pod \"ovn-operator-controller-manager-b6456fdb6-5zjf8\" (UID: \"42800642-ed16-4b36-8587-8d136d065283\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.735901 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24j66\" (UniqueName: \"kubernetes.io/projected/eb70d75b-9337-43a0-960b-ec19d05690bc-kube-api-access-24j66\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.735938 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.748461 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.803259 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.806135 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.813404 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4n7np" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.817250 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.825435 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.826705 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.830814 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-q742v" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.842203 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx9fg\" (UniqueName: \"kubernetes.io/projected/42800642-ed16-4b36-8587-8d136d065283-kube-api-access-zx9fg\") pod \"ovn-operator-controller-manager-b6456fdb6-5zjf8\" (UID: \"42800642-ed16-4b36-8587-8d136d065283\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.842524 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24j66\" (UniqueName: \"kubernetes.io/projected/eb70d75b-9337-43a0-960b-ec19d05690bc-kube-api-access-24j66\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.842571 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.842885 4867 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.842950 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert podName:eb70d75b-9337-43a0-960b-ec19d05690bc nodeName:}" failed. No retries permitted until 2025-12-12 07:04:42.34292824 +0000 UTC m=+969.914309579 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert") pod "openstack-baremetal-operator-controller-manager-5c747c4-kzx62" (UID: "eb70d75b-9337-43a0-960b-ec19d05690bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.851425 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.882054 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx9fg\" (UniqueName: \"kubernetes.io/projected/42800642-ed16-4b36-8587-8d136d065283-kube-api-access-zx9fg\") pod \"ovn-operator-controller-manager-b6456fdb6-5zjf8\" (UID: \"42800642-ed16-4b36-8587-8d136d065283\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.888941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24j66\" (UniqueName: \"kubernetes.io/projected/eb70d75b-9337-43a0-960b-ec19d05690bc-kube-api-access-24j66\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.906985 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.916301 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc"] Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.917675 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.929519 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-xn68q" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.951952 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnqk7\" (UniqueName: \"kubernetes.io/projected/86d48a0c-c93a-4090-8f13-f9d6ec122a68-kube-api-access-jnqk7\") pod \"placement-operator-controller-manager-78f8948974-vhfhv\" (UID: \"86d48a0c-c93a-4090-8f13-f9d6ec122a68\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.952044 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.952126 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s9b2\" (UniqueName: \"kubernetes.io/projected/b63f08d0-fff4-4852-9cd2-6467bf523366-kube-api-access-2s9b2\") pod \"swift-operator-controller-manager-9d58d64bc-r24lf\" (UID: \"b63f08d0-fff4-4852-9cd2-6467bf523366\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.952324 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: E1212 07:04:41.952364 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:42.952350487 +0000 UTC m=+970.523731746 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:41 crc kubenswrapper[4867]: I1212 07:04:41.960501 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.003741 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.006340 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.009259 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-p4szq" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.015268 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.024110 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.025974 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.031322 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.031791 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8z262" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.035691 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.064055 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnqk7\" (UniqueName: \"kubernetes.io/projected/86d48a0c-c93a-4090-8f13-f9d6ec122a68-kube-api-access-jnqk7\") pod \"placement-operator-controller-manager-78f8948974-vhfhv\" (UID: \"86d48a0c-c93a-4090-8f13-f9d6ec122a68\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.064413 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt9jf\" (UniqueName: \"kubernetes.io/projected/e17f7472-9f06-46f7-a911-96a88b93d30e-kube-api-access-pt9jf\") pod \"telemetry-operator-controller-manager-58d5ff84df-q84bc\" (UID: \"e17f7472-9f06-46f7-a911-96a88b93d30e\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.064898 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s9b2\" (UniqueName: \"kubernetes.io/projected/b63f08d0-fff4-4852-9cd2-6467bf523366-kube-api-access-2s9b2\") pod \"swift-operator-controller-manager-9d58d64bc-r24lf\" (UID: \"b63f08d0-fff4-4852-9cd2-6467bf523366\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.107344 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s9b2\" (UniqueName: \"kubernetes.io/projected/b63f08d0-fff4-4852-9cd2-6467bf523366-kube-api-access-2s9b2\") pod \"swift-operator-controller-manager-9d58d64bc-r24lf\" (UID: \"b63f08d0-fff4-4852-9cd2-6467bf523366\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.108766 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnqk7\" (UniqueName: \"kubernetes.io/projected/86d48a0c-c93a-4090-8f13-f9d6ec122a68-kube-api-access-jnqk7\") pod \"placement-operator-controller-manager-78f8948974-vhfhv\" (UID: \"86d48a0c-c93a-4090-8f13-f9d6ec122a68\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.118604 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.143545 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.143716 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.145487 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.145821 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.147462 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.147631 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7rv9v" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.154556 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.155822 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.162683 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-rsczz" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.162960 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.167930 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt9jf\" (UniqueName: \"kubernetes.io/projected/e17f7472-9f06-46f7-a911-96a88b93d30e-kube-api-access-pt9jf\") pod \"telemetry-operator-controller-manager-58d5ff84df-q84bc\" (UID: \"e17f7472-9f06-46f7-a911-96a88b93d30e\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.168044 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km9pz\" (UniqueName: \"kubernetes.io/projected/11ecafd2-343b-4ee3-9ffa-c6987230ccc3-kube-api-access-km9pz\") pod \"test-operator-controller-manager-5854674fcc-4cfxw\" (UID: \"11ecafd2-343b-4ee3-9ffa-c6987230ccc3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.168073 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zkjs\" (UniqueName: \"kubernetes.io/projected/a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04-kube-api-access-5zkjs\") pod \"watcher-operator-controller-manager-75944c9b7-vnbsx\" (UID: \"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.169929 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.190674 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.196860 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt9jf\" (UniqueName: \"kubernetes.io/projected/e17f7472-9f06-46f7-a911-96a88b93d30e-kube-api-access-pt9jf\") pod \"telemetry-operator-controller-manager-58d5ff84df-q84bc\" (UID: \"e17f7472-9f06-46f7-a911-96a88b93d30e\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.198399 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.271787 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.271858 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km9pz\" (UniqueName: \"kubernetes.io/projected/11ecafd2-343b-4ee3-9ffa-c6987230ccc3-kube-api-access-km9pz\") pod \"test-operator-controller-manager-5854674fcc-4cfxw\" (UID: \"11ecafd2-343b-4ee3-9ffa-c6987230ccc3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.271887 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zkjs\" (UniqueName: \"kubernetes.io/projected/a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04-kube-api-access-5zkjs\") pod \"watcher-operator-controller-manager-75944c9b7-vnbsx\" (UID: \"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.271920 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4bj\" (UniqueName: \"kubernetes.io/projected/08998d49-f3b6-4849-8182-8199f1f5c4c6-kube-api-access-mk4bj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7rkgr\" (UID: \"08998d49-f3b6-4849-8182-8199f1f5c4c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.278502 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7cmr\" (UniqueName: \"kubernetes.io/projected/9a77c831-df7a-4167-a8d4-5ca630bbefb0-kube-api-access-r7cmr\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.278582 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.295624 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km9pz\" (UniqueName: \"kubernetes.io/projected/11ecafd2-343b-4ee3-9ffa-c6987230ccc3-kube-api-access-km9pz\") pod \"test-operator-controller-manager-5854674fcc-4cfxw\" (UID: \"11ecafd2-343b-4ee3-9ffa-c6987230ccc3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.296479 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zkjs\" (UniqueName: \"kubernetes.io/projected/a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04-kube-api-access-5zkjs\") pod \"watcher-operator-controller-manager-75944c9b7-vnbsx\" (UID: \"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.352771 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.361606 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.380029 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7cmr\" (UniqueName: \"kubernetes.io/projected/9a77c831-df7a-4167-a8d4-5ca630bbefb0-kube-api-access-r7cmr\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.380072 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.380117 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.380179 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.380209 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4bj\" (UniqueName: \"kubernetes.io/projected/08998d49-f3b6-4849-8182-8199f1f5c4c6-kube-api-access-mk4bj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7rkgr\" (UID: \"08998d49-f3b6-4849-8182-8199f1f5c4c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.380820 4867 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.380862 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert podName:eb70d75b-9337-43a0-960b-ec19d05690bc nodeName:}" failed. No retries permitted until 2025-12-12 07:04:43.38084959 +0000 UTC m=+970.952230859 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert") pod "openstack-baremetal-operator-controller-manager-5c747c4-kzx62" (UID: "eb70d75b-9337-43a0-960b-ec19d05690bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.381137 4867 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.381167 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:42.881158087 +0000 UTC m=+970.452539356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "metrics-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.381200 4867 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.381218 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:42.881212979 +0000 UTC m=+970.452594248 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.400906 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4bj\" (UniqueName: \"kubernetes.io/projected/08998d49-f3b6-4849-8182-8199f1f5c4c6-kube-api-access-mk4bj\") pod \"rabbitmq-cluster-operator-manager-668c99d594-7rkgr\" (UID: \"08998d49-f3b6-4849-8182-8199f1f5c4c6\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.402332 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7cmr\" (UniqueName: \"kubernetes.io/projected/9a77c831-df7a-4167-a8d4-5ca630bbefb0-kube-api-access-r7cmr\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.408714 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.457071 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.611525 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.890316 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.890416 4867 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.890803 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:43.890784017 +0000 UTC m=+971.462165286 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.890824 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.890914 4867 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.890969 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:43.890935101 +0000 UTC m=+971.462316380 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "metrics-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.936685 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.958752 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" event={"ID":"eab0a735-13ed-4d91-a4c5-d3c09d99139d","Type":"ContainerStarted","Data":"cc5d04316cd8517139a50bc3bbed094afd38c069b402eb69ed66cd6d08e91ec6"} Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.964938 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx"] Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.967099 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" event={"ID":"8731ddb0-38b0-4de8-895a-c880c8899d26","Type":"ContainerStarted","Data":"c3a28150e843e59755270477d34371203326645af496dba05f4f3e654902b97c"} Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.969012 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" event={"ID":"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f","Type":"ContainerStarted","Data":"e251833e13e29076286190419b1b1b1cd2e7b83aee869daebf2166e03085a5cb"} Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.978524 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9"] Dec 12 07:04:42 crc kubenswrapper[4867]: W1212 07:04:42.979555 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c8a7928_c0c9_4516_8e34_273afc40ccae.slice/crio-96da0dd711cc1339dd7ca247ecaf774dbedae463380071f8479371949de03a59 WatchSource:0}: Error finding container 96da0dd711cc1339dd7ca247ecaf774dbedae463380071f8479371949de03a59: Status 404 returned error can't find the container with id 96da0dd711cc1339dd7ca247ecaf774dbedae463380071f8479371949de03a59 Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.979811 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" event={"ID":"d774c064-6d06-4d04-8c11-7008260c0044","Type":"ContainerStarted","Data":"a3ad2426b548082952a31122dcd3e1ae323d4c176b15aee505836f7c98f4b0a0"} Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.994064 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.996507 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: E1212 07:04:42.996560 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:44.996543894 +0000 UTC m=+972.567925163 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:42 crc kubenswrapper[4867]: I1212 07:04:42.996589 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.007758 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.050892 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.054953 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.064246 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj"] Dec 12 07:04:43 crc kubenswrapper[4867]: W1212 07:04:43.072019 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42800642_ed16_4b36_8587_8d136d065283.slice/crio-b71d705816f90a76398c6b1443e3f94b7dcca81ce9fecc2611500a2c8ae52be4 WatchSource:0}: Error finding container b71d705816f90a76398c6b1443e3f94b7dcca81ce9fecc2611500a2c8ae52be4: Status 404 returned error can't find the container with id b71d705816f90a76398c6b1443e3f94b7dcca81ce9fecc2611500a2c8ae52be4 Dec 12 07:04:43 crc kubenswrapper[4867]: W1212 07:04:43.073569 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf70067b7_bc5e_4f40_9b90_7601ac37aa89.slice/crio-49a22428647e672d20775b8af85f86c2a8485a83aefffdc0419bcfb327143554 WatchSource:0}: Error finding container 49a22428647e672d20775b8af85f86c2a8485a83aefffdc0419bcfb327143554: Status 404 returned error can't find the container with id 49a22428647e672d20775b8af85f86c2a8485a83aefffdc0419bcfb327143554 Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.074634 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.080914 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.087143 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk"] Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.091092 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whgrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-c8dlk_openstack-operators(b29bfc3e-61ea-443a-af46-13583b7d619f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.093962 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whgrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-c8dlk_openstack-operators(b29bfc3e-61ea-443a-af46-13583b7d619f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.095079 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" podUID="b29bfc3e-61ea-443a-af46-13583b7d619f" Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.274862 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.283492 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw"] Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.288695 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf"] Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.294039 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mk4bj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-7rkgr_openstack-operators(08998d49-f3b6-4849-8182-8199f1f5c4c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.294139 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pt9jf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-q84bc_openstack-operators(e17f7472-9f06-46f7-a911-96a88b93d30e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.294346 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnqk7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-vhfhv_openstack-operators(86d48a0c-c93a-4090-8f13-f9d6ec122a68): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.294963 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv"] Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.295139 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" podUID="08998d49-f3b6-4849-8182-8199f1f5c4c6" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.297814 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pt9jf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-q84bc_openstack-operators(e17f7472-9f06-46f7-a911-96a88b93d30e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.297871 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnqk7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-vhfhv_openstack-operators(86d48a0c-c93a-4090-8f13-f9d6ec122a68): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: W1212 07:04:43.299131 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda01f2b25_dab2_4d4d_8d6a_ba3249ec2f04.slice/crio-0c52acb5d073132e8d680d94308374977ba306130abca1dcf6eb3782006df373 WatchSource:0}: Error finding container 0c52acb5d073132e8d680d94308374977ba306130abca1dcf6eb3782006df373: Status 404 returned error can't find the container with id 0c52acb5d073132e8d680d94308374977ba306130abca1dcf6eb3782006df373 Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.299240 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" podUID="e17f7472-9f06-46f7-a911-96a88b93d30e" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.299925 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" podUID="86d48a0c-c93a-4090-8f13-f9d6ec122a68" Dec 12 07:04:43 crc kubenswrapper[4867]: W1212 07:04:43.301624 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11ecafd2_343b_4ee3_9ffa_c6987230ccc3.slice/crio-5165f0abb89236a775262614157134a11cd473c3a5e1b3fef4db88cc2cf846b0 WatchSource:0}: Error finding container 5165f0abb89236a775262614157134a11cd473c3a5e1b3fef4db88cc2cf846b0: Status 404 returned error can't find the container with id 5165f0abb89236a775262614157134a11cd473c3a5e1b3fef4db88cc2cf846b0 Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.303034 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx"] Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.306576 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-km9pz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4cfxw_openstack-operators(11ecafd2-343b-4ee3-9ffa-c6987230ccc3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.308355 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr"] Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.309920 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-km9pz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-4cfxw_openstack-operators(11ecafd2-343b-4ee3-9ffa-c6987230ccc3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.311116 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" podUID="11ecafd2-343b-4ee3-9ffa-c6987230ccc3" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.321294 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5zkjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-vnbsx_openstack-operators(a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.323900 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5zkjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-vnbsx_openstack-operators(a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.324958 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" podUID="a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04" Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.402405 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.402720 4867 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.402772 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert podName:eb70d75b-9337-43a0-960b-ec19d05690bc nodeName:}" failed. No retries permitted until 2025-12-12 07:04:45.402757465 +0000 UTC m=+972.974138734 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert") pod "openstack-baremetal-operator-controller-manager-5c747c4-kzx62" (UID: "eb70d75b-9337-43a0-960b-ec19d05690bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.925781 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.925923 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.926097 4867 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.926159 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:45.926140955 +0000 UTC m=+973.497522224 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "metrics-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.926619 4867 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.926673 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:45.926661858 +0000 UTC m=+973.498043127 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "webhook-server-cert" not found Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.997080 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" event={"ID":"08998d49-f3b6-4849-8182-8199f1f5c4c6","Type":"ContainerStarted","Data":"126f06b505cf61391228c981d5e3a44bf329d7004421b84b0f58c6b6ca42e05f"} Dec 12 07:04:43 crc kubenswrapper[4867]: I1212 07:04:43.998065 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" event={"ID":"b29bfc3e-61ea-443a-af46-13583b7d619f","Type":"ContainerStarted","Data":"102db1bfe7aab7333aeed7194b49afdc143362e6896401b88d3410371a0815ed"} Dec 12 07:04:43 crc kubenswrapper[4867]: E1212 07:04:43.999506 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" podUID="08998d49-f3b6-4849-8182-8199f1f5c4c6" Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.000912 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" event={"ID":"11ecafd2-343b-4ee3-9ffa-c6987230ccc3","Type":"ContainerStarted","Data":"5165f0abb89236a775262614157134a11cd473c3a5e1b3fef4db88cc2cf846b0"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.002736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" event={"ID":"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04","Type":"ContainerStarted","Data":"0c52acb5d073132e8d680d94308374977ba306130abca1dcf6eb3782006df373"} Dec 12 07:04:44 crc kubenswrapper[4867]: E1212 07:04:44.003634 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" podUID="b29bfc3e-61ea-443a-af46-13583b7d619f" Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.003915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" event={"ID":"09531d04-5bd7-4457-af43-296c2e76cdfb","Type":"ContainerStarted","Data":"a26e4459d47d60a5a54ee3cb65655764df46056668156130672dd44b4de20dab"} Dec 12 07:04:44 crc kubenswrapper[4867]: E1212 07:04:44.004807 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" podUID="11ecafd2-343b-4ee3-9ffa-c6987230ccc3" Dec 12 07:04:44 crc kubenswrapper[4867]: E1212 07:04:44.004926 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" podUID="a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04" Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.005726 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" event={"ID":"86d48a0c-c93a-4090-8f13-f9d6ec122a68","Type":"ContainerStarted","Data":"b21ec720db0e65d4547b3701b95daea4aa284c20ceeb7d83471e29e8879dfb54"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.009770 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" event={"ID":"9c8a7928-c0c9-4516-8e34-273afc40ccae","Type":"ContainerStarted","Data":"96da0dd711cc1339dd7ca247ecaf774dbedae463380071f8479371949de03a59"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.013270 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" event={"ID":"528acf70-c2ad-4968-88ad-24531e8f5d12","Type":"ContainerStarted","Data":"1a6d56900ef6454f0f639c6e5c52eb471b93af40a32d7e8e0f8e44c2bca9c268"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.016175 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" event={"ID":"f70067b7-bc5e-4f40-9b90-7601ac37aa89","Type":"ContainerStarted","Data":"49a22428647e672d20775b8af85f86c2a8485a83aefffdc0419bcfb327143554"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.018931 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" event={"ID":"e17f7472-9f06-46f7-a911-96a88b93d30e","Type":"ContainerStarted","Data":"d8d44e677b72b3b48025196948a4d95a04e3d5a49cf313847c8e4db5ecfdc827"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.021417 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" event={"ID":"42800642-ed16-4b36-8587-8d136d065283","Type":"ContainerStarted","Data":"b71d705816f90a76398c6b1443e3f94b7dcca81ce9fecc2611500a2c8ae52be4"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.023251 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" event={"ID":"b63f08d0-fff4-4852-9cd2-6467bf523366","Type":"ContainerStarted","Data":"96fd4e5f2bb48dc05666460be6bbbbee9272f5cea028e5d186665d36fa4ea6ec"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.025331 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" event={"ID":"09951d5e-0bdd-4a76-b433-880fa1bb7abd","Type":"ContainerStarted","Data":"1c85e6fe3c4fdcdf146c772df11e22a1af8e75f42e289db9e532ee496fb249f6"} Dec 12 07:04:44 crc kubenswrapper[4867]: E1212 07:04:44.028149 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" podUID="86d48a0c-c93a-4090-8f13-f9d6ec122a68" Dec 12 07:04:44 crc kubenswrapper[4867]: E1212 07:04:44.030752 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" podUID="e17f7472-9f06-46f7-a911-96a88b93d30e" Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.030992 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" event={"ID":"93fd7786-fb3e-45aa-92d2-6b6fa38daf56","Type":"ContainerStarted","Data":"2f43529f02a72955d965a0b327372ace79e8722ff83b24c19d3302384688d813"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.035575 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" event={"ID":"5a7bb06a-0518-4384-b694-b73194df4a05","Type":"ContainerStarted","Data":"96f382563580ee367c803a791dee2cbb48ea6a4d094a7e27940892e55eb2a4d1"} Dec 12 07:04:44 crc kubenswrapper[4867]: I1212 07:04:44.039651 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" event={"ID":"eacc6270-8929-44a1-a971-f2a6b7582103","Type":"ContainerStarted","Data":"e24d7dca8c993c0c5680648af9496945e16bb67b91d24293e287b4a3a162dfc8"} Dec 12 07:04:45 crc kubenswrapper[4867]: I1212 07:04:45.043665 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.043809 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.043882 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:49.043859921 +0000 UTC m=+976.615241190 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.052563 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" podUID="08998d49-f3b6-4849-8182-8199f1f5c4c6" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.054029 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" podUID="11ecafd2-343b-4ee3-9ffa-c6987230ccc3" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.054130 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" podUID="b29bfc3e-61ea-443a-af46-13583b7d619f" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.054405 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" podUID="a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.057080 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" podUID="e17f7472-9f06-46f7-a911-96a88b93d30e" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.057497 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" podUID="86d48a0c-c93a-4090-8f13-f9d6ec122a68" Dec 12 07:04:45 crc kubenswrapper[4867]: I1212 07:04:45.456157 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.456598 4867 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.456688 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert podName:eb70d75b-9337-43a0-960b-ec19d05690bc nodeName:}" failed. No retries permitted until 2025-12-12 07:04:49.456664395 +0000 UTC m=+977.028045664 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert") pod "openstack-baremetal-operator-controller-manager-5c747c4-kzx62" (UID: "eb70d75b-9337-43a0-960b-ec19d05690bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: I1212 07:04:45.973557 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.974000 4867 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.974160 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:49.974145638 +0000 UTC m=+977.545526907 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "metrics-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: I1212 07:04:45.974514 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.975365 4867 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 12 07:04:45 crc kubenswrapper[4867]: E1212 07:04:45.975410 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:49.975398659 +0000 UTC m=+977.546780008 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "webhook-server-cert" not found Dec 12 07:04:49 crc kubenswrapper[4867]: I1212 07:04:49.126143 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:49 crc kubenswrapper[4867]: E1212 07:04:49.126338 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:49 crc kubenswrapper[4867]: E1212 07:04:49.126418 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:57.126396335 +0000 UTC m=+984.697777654 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:49 crc kubenswrapper[4867]: I1212 07:04:49.531514 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:49 crc kubenswrapper[4867]: E1212 07:04:49.531761 4867 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:49 crc kubenswrapper[4867]: E1212 07:04:49.531809 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert podName:eb70d75b-9337-43a0-960b-ec19d05690bc nodeName:}" failed. No retries permitted until 2025-12-12 07:04:57.531794643 +0000 UTC m=+985.103175912 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert") pod "openstack-baremetal-operator-controller-manager-5c747c4-kzx62" (UID: "eb70d75b-9337-43a0-960b-ec19d05690bc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 12 07:04:50 crc kubenswrapper[4867]: I1212 07:04:50.037468 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:50 crc kubenswrapper[4867]: E1212 07:04:50.037658 4867 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 12 07:04:50 crc kubenswrapper[4867]: I1212 07:04:50.037904 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:50 crc kubenswrapper[4867]: E1212 07:04:50.037947 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:58.037928599 +0000 UTC m=+985.609309868 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "webhook-server-cert" not found Dec 12 07:04:50 crc kubenswrapper[4867]: E1212 07:04:50.038066 4867 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 12 07:04:50 crc kubenswrapper[4867]: E1212 07:04:50.038121 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs podName:9a77c831-df7a-4167-a8d4-5ca630bbefb0 nodeName:}" failed. No retries permitted until 2025-12-12 07:04:58.038106133 +0000 UTC m=+985.609487402 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs") pod "openstack-operator-controller-manager-5b74fbd87-zjj58" (UID: "9a77c831-df7a-4167-a8d4-5ca630bbefb0") : secret "metrics-server-cert" not found Dec 12 07:04:57 crc kubenswrapper[4867]: I1212 07:04:57.167946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:04:57 crc kubenswrapper[4867]: E1212 07:04:57.168185 4867 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:57 crc kubenswrapper[4867]: E1212 07:04:57.168696 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert podName:6c74d9ef-6d91-4c62-bfe4-e4811b78adf6 nodeName:}" failed. No retries permitted until 2025-12-12 07:05:13.168672219 +0000 UTC m=+1000.740053488 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert") pod "infra-operator-controller-manager-78d48bff9d-jqlwv" (UID: "6c74d9ef-6d91-4c62-bfe4-e4811b78adf6") : secret "infra-operator-webhook-server-cert" not found Dec 12 07:04:57 crc kubenswrapper[4867]: I1212 07:04:57.573495 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:57 crc kubenswrapper[4867]: I1212 07:04:57.579589 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb70d75b-9337-43a0-960b-ec19d05690bc-cert\") pod \"openstack-baremetal-operator-controller-manager-5c747c4-kzx62\" (UID: \"eb70d75b-9337-43a0-960b-ec19d05690bc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:57 crc kubenswrapper[4867]: I1212 07:04:57.700423 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.080014 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.080117 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.083413 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-webhook-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.083413 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9a77c831-df7a-4167-a8d4-5ca630bbefb0-metrics-certs\") pod \"openstack-operator-controller-manager-5b74fbd87-zjj58\" (UID: \"9a77c831-df7a-4167-a8d4-5ca630bbefb0\") " pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.126138 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.988791 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:04:58 crc kubenswrapper[4867]: I1212 07:04:58.988878 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:05:04 crc kubenswrapper[4867]: E1212 07:05:04.174032 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 12 07:05:04 crc kubenswrapper[4867]: E1212 07:05:04.174621 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2s9b2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-r24lf_openstack-operators(b63f08d0-fff4-4852-9cd2-6467bf523366): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:05:04 crc kubenswrapper[4867]: E1212 07:05:04.793532 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 12 07:05:04 crc kubenswrapper[4867]: E1212 07:05:04.793698 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dsgzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-62qkh_openstack-operators(9c8a7928-c0c9-4516-8e34-273afc40ccae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:05:05 crc kubenswrapper[4867]: E1212 07:05:05.476236 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 12 07:05:05 crc kubenswrapper[4867]: E1212 07:05:05.476752 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8dmwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-6drsk_openstack-operators(09951d5e-0bdd-4a76-b433-880fa1bb7abd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.026548 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62"] Dec 12 07:05:06 crc kubenswrapper[4867]: W1212 07:05:06.064474 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb70d75b_9337_43a0_960b_ec19d05690bc.slice/crio-e9bf11693fd629cfba855167806ef63e16a7a95cca1a7f49fa048e45a8383b05 WatchSource:0}: Error finding container e9bf11693fd629cfba855167806ef63e16a7a95cca1a7f49fa048e45a8383b05: Status 404 returned error can't find the container with id e9bf11693fd629cfba855167806ef63e16a7a95cca1a7f49fa048e45a8383b05 Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.140024 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58"] Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.285298 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" event={"ID":"42800642-ed16-4b36-8587-8d136d065283","Type":"ContainerStarted","Data":"728e4683068de3427478a05c8ed806cd5413bf5d36ca4f37c5ec1affaddf67f4"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.292248 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" event={"ID":"eb70d75b-9337-43a0-960b-ec19d05690bc","Type":"ContainerStarted","Data":"e9bf11693fd629cfba855167806ef63e16a7a95cca1a7f49fa048e45a8383b05"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.354529 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" event={"ID":"5a7bb06a-0518-4384-b694-b73194df4a05","Type":"ContainerStarted","Data":"0f3254efc6e90c4f1875b129de5fd586a9b7c49f78056bf0da110da486ad7986"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.357954 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" event={"ID":"528acf70-c2ad-4968-88ad-24531e8f5d12","Type":"ContainerStarted","Data":"875e12dfdf7d768eb548bea968ee8bf361d6945b68e2b81950a32d2055d906e2"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.376499 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" event={"ID":"eab0a735-13ed-4d91-a4c5-d3c09d99139d","Type":"ContainerStarted","Data":"12e2f13b98c0b76324ede2d771960ddf7fea02809a45c1ebdb6049a868e9fea5"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.379739 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" event={"ID":"93fd7786-fb3e-45aa-92d2-6b6fa38daf56","Type":"ContainerStarted","Data":"08bd812627ec0bf2216c21ce8a9292329bbda377b469463cd447bb05988e7329"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.400962 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" event={"ID":"8731ddb0-38b0-4de8-895a-c880c8899d26","Type":"ContainerStarted","Data":"10842c4166261e45fe71912dcbab340c65dcf54196ba1eb1e4f0270b0ef58325"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.408694 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" event={"ID":"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f","Type":"ContainerStarted","Data":"393fe8d31ba8594f998340b3068499b79057bdebbd695d3a3be921c06a3300f7"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.417107 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" event={"ID":"f70067b7-bc5e-4f40-9b90-7601ac37aa89","Type":"ContainerStarted","Data":"cc80cf39a283f92e701c02d4f8f0206160b618e7d006ddc7f461a5eeff62bb24"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.429019 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" event={"ID":"09531d04-5bd7-4457-af43-296c2e76cdfb","Type":"ContainerStarted","Data":"3a9d6a02d0a1f0a10105c102d7c6aeb6fbfa7999abc1a3c04544cda615df77ac"} Dec 12 07:05:06 crc kubenswrapper[4867]: I1212 07:05:06.450904 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" event={"ID":"eacc6270-8929-44a1-a971-f2a6b7582103","Type":"ContainerStarted","Data":"cfbc56b184e82aca869bc9257d94ede4f1cf39ead64b18d33e5fc41e0f859225"} Dec 12 07:05:07 crc kubenswrapper[4867]: W1212 07:05:07.162983 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a77c831_df7a_4167_a8d4_5ca630bbefb0.slice/crio-89110938d172248cac39dd192c6aa9647c9a71c967d369a89cd098d636fd96d7 WatchSource:0}: Error finding container 89110938d172248cac39dd192c6aa9647c9a71c967d369a89cd098d636fd96d7: Status 404 returned error can't find the container with id 89110938d172248cac39dd192c6aa9647c9a71c967d369a89cd098d636fd96d7 Dec 12 07:05:07 crc kubenswrapper[4867]: I1212 07:05:07.473163 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" event={"ID":"d774c064-6d06-4d04-8c11-7008260c0044","Type":"ContainerStarted","Data":"5386a32b7dc5f80b3452a8ce4d218ca6b8b1539a4a9e7c506bd1b8516b5373e3"} Dec 12 07:05:07 crc kubenswrapper[4867]: I1212 07:05:07.474484 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" event={"ID":"9a77c831-df7a-4167-a8d4-5ca630bbefb0","Type":"ContainerStarted","Data":"89110938d172248cac39dd192c6aa9647c9a71c967d369a89cd098d636fd96d7"} Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.236166 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.248388 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6c74d9ef-6d91-4c62-bfe4-e4811b78adf6-cert\") pod \"infra-operator-controller-manager-78d48bff9d-jqlwv\" (UID: \"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.288515 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.516526 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" event={"ID":"9a77c831-df7a-4167-a8d4-5ca630bbefb0","Type":"ContainerStarted","Data":"23ca4dadcb77bf3ec960e1478b3be96dee5140ea0fb702a01e8d02f780c53fd0"} Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.516738 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:05:13 crc kubenswrapper[4867]: I1212 07:05:13.563401 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" podStartSLOduration=32.56338028 podStartE2EDuration="32.56338028s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:05:13.558345227 +0000 UTC m=+1001.129726506" watchObservedRunningTime="2025-12-12 07:05:13.56338028 +0000 UTC m=+1001.134761549" Dec 12 07:05:14 crc kubenswrapper[4867]: E1212 07:05:14.106013 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" podUID="b63f08d0-fff4-4852-9cd2-6467bf523366" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.208621 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv"] Dec 12 07:05:14 crc kubenswrapper[4867]: W1212 07:05:14.216126 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c74d9ef_6d91_4c62_bfe4_e4811b78adf6.slice/crio-ad4bd0e87452e46ff59baa488b2a7d1982b8aa06f89068b8bbf217f74fde8d36 WatchSource:0}: Error finding container ad4bd0e87452e46ff59baa488b2a7d1982b8aa06f89068b8bbf217f74fde8d36: Status 404 returned error can't find the container with id ad4bd0e87452e46ff59baa488b2a7d1982b8aa06f89068b8bbf217f74fde8d36 Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.261458 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.577416 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" event={"ID":"eacc6270-8929-44a1-a971-f2a6b7582103","Type":"ContainerStarted","Data":"fb743223fffe72459fbb259df11737cb59f919cf6ef58330551397697d14036e"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.578059 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.581600 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.586890 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" event={"ID":"93fd7786-fb3e-45aa-92d2-6b6fa38daf56","Type":"ContainerStarted","Data":"5ab006808821b8d15b561f708d0b350bb7dfb6238fdfa05b31e6b8617fd4f923"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.587675 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.607170 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.613538 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-rfkwx" podStartSLOduration=2.77827154 podStartE2EDuration="33.613514605s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.962929443 +0000 UTC m=+970.534310712" lastFinishedPulling="2025-12-12 07:05:13.798172508 +0000 UTC m=+1001.369553777" observedRunningTime="2025-12-12 07:05:14.607546929 +0000 UTC m=+1002.178928198" watchObservedRunningTime="2025-12-12 07:05:14.613514605 +0000 UTC m=+1002.184895884" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.635533 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" event={"ID":"eb70d75b-9337-43a0-960b-ec19d05690bc","Type":"ContainerStarted","Data":"abd54ecde93327082e373dd1b6598a21906c0b4cf8e25adffefa85a73656142c"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.692906 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" event={"ID":"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04","Type":"ContainerStarted","Data":"ae830ed24186f544f8babd1e0f0a16da83d8604c7e4740d260f415ba57bd10de"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.699129 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-z9jcv" podStartSLOduration=2.802207952 podStartE2EDuration="33.699088172s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.955767775 +0000 UTC m=+970.527149044" lastFinishedPulling="2025-12-12 07:05:13.852647995 +0000 UTC m=+1001.424029264" observedRunningTime="2025-12-12 07:05:14.692888711 +0000 UTC m=+1002.264269990" watchObservedRunningTime="2025-12-12 07:05:14.699088172 +0000 UTC m=+1002.270469441" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.719501 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" event={"ID":"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6","Type":"ContainerStarted","Data":"ad4bd0e87452e46ff59baa488b2a7d1982b8aa06f89068b8bbf217f74fde8d36"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.781642 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" event={"ID":"e17f7472-9f06-46f7-a911-96a88b93d30e","Type":"ContainerStarted","Data":"0380a07e768b72d50eaba184b41dc2d853856c963caf963f7b28e0c6532cf874"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.781697 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" event={"ID":"e17f7472-9f06-46f7-a911-96a88b93d30e","Type":"ContainerStarted","Data":"0e3a2ee1422c40e324ea488f697d7a2a28eff350aaccbfbabeeaea901df015e4"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.782804 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.820172 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" podStartSLOduration=4.669997187 podStartE2EDuration="33.820151454s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.293976003 +0000 UTC m=+970.865357272" lastFinishedPulling="2025-12-12 07:05:12.44413028 +0000 UTC m=+1000.015511539" observedRunningTime="2025-12-12 07:05:14.819617332 +0000 UTC m=+1002.390998601" watchObservedRunningTime="2025-12-12 07:05:14.820151454 +0000 UTC m=+1002.391532723" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.863897 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" event={"ID":"86d48a0c-c93a-4090-8f13-f9d6ec122a68","Type":"ContainerStarted","Data":"b418933acdf45d4d063065bb6ce98aac7791fbdbff85d3073ea6f3f5f72e9710"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.897353 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" event={"ID":"b63f08d0-fff4-4852-9cd2-6467bf523366","Type":"ContainerStarted","Data":"7030aa8889e63e24eb81b46c2b43bc8e68ce5969052c76bbec8456e521ee9cad"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.900127 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" event={"ID":"08998d49-f3b6-4849-8182-8199f1f5c4c6","Type":"ContainerStarted","Data":"3164313e91fc05b165331a2bbbf169b83f2b4ce89da6db45e5f3c3661e446d07"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.930476 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" event={"ID":"b29bfc3e-61ea-443a-af46-13583b7d619f","Type":"ContainerStarted","Data":"0c58ed399185554d595a27257707278a624c913ef0552a52db839d0bea1bacd3"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.932497 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" event={"ID":"f70067b7-bc5e-4f40-9b90-7601ac37aa89","Type":"ContainerStarted","Data":"acbb2d3939b130f096627df81cca5ceef2fb5cbbe12b108e8bcdfcc31e3532b5"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.934846 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.936642 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.962079 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" event={"ID":"42800642-ed16-4b36-8587-8d136d065283","Type":"ContainerStarted","Data":"aa9e2cc45e19bc99c39cac56b927708926131270fcb3fd8d9b9b89af56e652c5"} Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.963066 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.978453 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-7rkgr" podStartSLOduration=4.719186989 podStartE2EDuration="33.978422236s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.293763948 +0000 UTC m=+970.865145217" lastFinishedPulling="2025-12-12 07:05:12.552999195 +0000 UTC m=+1000.124380464" observedRunningTime="2025-12-12 07:05:14.961630356 +0000 UTC m=+1002.533011635" watchObservedRunningTime="2025-12-12 07:05:14.978422236 +0000 UTC m=+1002.549803515" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.984858 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" Dec 12 07:05:14 crc kubenswrapper[4867]: I1212 07:05:14.988287 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" event={"ID":"11ecafd2-343b-4ee3-9ffa-c6987230ccc3","Type":"ContainerStarted","Data":"8f85fc3ac13f5489193820a6c5cd0efb084a3fc11a37ecf2bd0a3b3db6944b2d"} Dec 12 07:05:15 crc kubenswrapper[4867]: I1212 07:05:15.045813 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-frb6j" podStartSLOduration=3.216585043 podStartE2EDuration="34.045787519s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.079050636 +0000 UTC m=+970.650431905" lastFinishedPulling="2025-12-12 07:05:13.908253112 +0000 UTC m=+1001.479634381" observedRunningTime="2025-12-12 07:05:15.031150781 +0000 UTC m=+1002.602532050" watchObservedRunningTime="2025-12-12 07:05:15.045787519 +0000 UTC m=+1002.617168788" Dec 12 07:05:15 crc kubenswrapper[4867]: E1212 07:05:15.580342 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" podUID="9c8a7928-c0c9-4516-8e34-273afc40ccae" Dec 12 07:05:15 crc kubenswrapper[4867]: E1212 07:05:15.758705 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" podUID="09951d5e-0bdd-4a76-b433-880fa1bb7abd" Dec 12 07:05:15 crc kubenswrapper[4867]: I1212 07:05:15.997186 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" event={"ID":"b29bfc3e-61ea-443a-af46-13583b7d619f","Type":"ContainerStarted","Data":"81ee58e915d45cbdf8128e6cf6b7bde3a6d95c0afa652cef928610e3a6a986c6"} Dec 12 07:05:15 crc kubenswrapper[4867]: I1212 07:05:15.997295 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.007634 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" event={"ID":"b63f08d0-fff4-4852-9cd2-6467bf523366","Type":"ContainerStarted","Data":"fef28e541bcd58980990479c08b2fbeb3881d02d0ad8c9bc9ed84042d5b1206c"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.007784 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.016633 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" event={"ID":"9c8a7928-c0c9-4516-8e34-273afc40ccae","Type":"ContainerStarted","Data":"75ab45e2c7ecb75f6a7491f370c335c10facf585508430a343d537d26c65b160"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.020386 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5zjf8" podStartSLOduration=4.296907881 podStartE2EDuration="35.0203718s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.073815336 +0000 UTC m=+970.645196605" lastFinishedPulling="2025-12-12 07:05:13.797279255 +0000 UTC m=+1001.368660524" observedRunningTime="2025-12-12 07:05:15.10367441 +0000 UTC m=+1002.675055679" watchObservedRunningTime="2025-12-12 07:05:16.0203718 +0000 UTC m=+1003.591753069" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.026438 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" event={"ID":"eab0a735-13ed-4d91-a4c5-d3c09d99139d","Type":"ContainerStarted","Data":"880314174a6e7f47ec341753d4cf6139887b0c0e4d36b634d706d7bc3efc5ef1"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.027373 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.031784 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.042161 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" event={"ID":"a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04","Type":"ContainerStarted","Data":"8f2783bdf6f8f909fcfe44907767f791a4e60f5337aeadb59dcd429033bd09a6"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.042276 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.048144 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" podStartSLOduration=5.587670971 podStartE2EDuration="35.048128588s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.09097134 +0000 UTC m=+970.662352609" lastFinishedPulling="2025-12-12 07:05:12.551428957 +0000 UTC m=+1000.122810226" observedRunningTime="2025-12-12 07:05:16.027421262 +0000 UTC m=+1003.598802531" watchObservedRunningTime="2025-12-12 07:05:16.048128588 +0000 UTC m=+1003.619509857" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.053184 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" event={"ID":"09531d04-5bd7-4457-af43-296c2e76cdfb","Type":"ContainerStarted","Data":"3d51a5071f4f153de1a24ce28f15e5d26ca6ebeb000aec3ef4530812a2adec87"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.054206 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.057348 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.066561 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" event={"ID":"d774c064-6d06-4d04-8c11-7008260c0044","Type":"ContainerStarted","Data":"642f0124da6fa926afc3da485c1c83d75308e7f73afad722f9cff765829242ed"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.067488 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.073417 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-2hqbr" podStartSLOduration=3.01268396 podStartE2EDuration="35.073403884s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.940483487 +0000 UTC m=+970.511864756" lastFinishedPulling="2025-12-12 07:05:15.001203401 +0000 UTC m=+1002.572584680" observedRunningTime="2025-12-12 07:05:16.073184148 +0000 UTC m=+1003.644565417" watchObservedRunningTime="2025-12-12 07:05:16.073403884 +0000 UTC m=+1003.644785153" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.079533 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.080355 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" event={"ID":"2962c7e5-a1d8-451b-9d65-f4e6e8295e2f","Type":"ContainerStarted","Data":"3d9bca8e6050c2f170b476df7fb40832254474cbaf21be5aaf4c54e69e09d6b2"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.081176 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.083903 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.084937 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" event={"ID":"eb70d75b-9337-43a0-960b-ec19d05690bc","Type":"ContainerStarted","Data":"93e909ec3e37f60981784bda9be42f41db460db9b6c8bf16fd7c119859e74404"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.085492 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.087821 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" event={"ID":"5a7bb06a-0518-4384-b694-b73194df4a05","Type":"ContainerStarted","Data":"cad4c4949791000c95b17c22c14d79d2b0388e588fd57884dcae0e26df532c1f"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.088529 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.090501 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.090712 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" event={"ID":"528acf70-c2ad-4968-88ad-24531e8f5d12","Type":"ContainerStarted","Data":"ee7458400f69ac7c71d6c7f0d5f5bd9514122707c22cb3bb2bdf3210961daef6"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.091370 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.096982 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" event={"ID":"8731ddb0-38b0-4de8-895a-c880c8899d26","Type":"ContainerStarted","Data":"4f7540226fa9b01340c25111e8e93dfd0ade72747de65dfb1228ce68d88ad540"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.097977 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.100957 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.101810 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" podStartSLOduration=2.7639949059999998 podStartE2EDuration="35.101800656s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.29306583 +0000 UTC m=+970.864447099" lastFinishedPulling="2025-12-12 07:05:15.63087158 +0000 UTC m=+1003.202252849" observedRunningTime="2025-12-12 07:05:16.093624947 +0000 UTC m=+1003.665006216" watchObservedRunningTime="2025-12-12 07:05:16.101800656 +0000 UTC m=+1003.673181925" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.109016 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.110367 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" event={"ID":"11ecafd2-343b-4ee3-9ffa-c6987230ccc3","Type":"ContainerStarted","Data":"425bdf5a27eba6b18660d764e79d58fc7b66efde71a1b7f31df5884287b61358"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.110946 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.120357 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-fhp5f" podStartSLOduration=3.872527441 podStartE2EDuration="35.120338918s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.082822608 +0000 UTC m=+970.654203877" lastFinishedPulling="2025-12-12 07:05:14.330634085 +0000 UTC m=+1001.902015354" observedRunningTime="2025-12-12 07:05:16.11629364 +0000 UTC m=+1003.687674909" watchObservedRunningTime="2025-12-12 07:05:16.120338918 +0000 UTC m=+1003.691720187" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.121273 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" event={"ID":"86d48a0c-c93a-4090-8f13-f9d6ec122a68","Type":"ContainerStarted","Data":"b48225a27b4b90d237a4412314e4489af68bb1dc13a3d8321ed5ff50e5baa1b9"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.121950 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.124971 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" event={"ID":"09951d5e-0bdd-4a76-b433-880fa1bb7abd","Type":"ContainerStarted","Data":"cbf833bdb73563829eaf36088ae45ff322dddd1b73ecc91db9a2a2abe27ad1f1"} Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.227671 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" podStartSLOduration=28.582977472 podStartE2EDuration="35.227654286s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:05:06.074445743 +0000 UTC m=+993.645827012" lastFinishedPulling="2025-12-12 07:05:12.719122557 +0000 UTC m=+1000.290503826" observedRunningTime="2025-12-12 07:05:16.225862952 +0000 UTC m=+1003.797244221" watchObservedRunningTime="2025-12-12 07:05:16.227654286 +0000 UTC m=+1003.799035565" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.256746 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-5mwnj" podStartSLOduration=4.600533319 podStartE2EDuration="36.256726006s" podCreationTimestamp="2025-12-12 07:04:40 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.083300351 +0000 UTC m=+970.654681620" lastFinishedPulling="2025-12-12 07:05:14.739493038 +0000 UTC m=+1002.310874307" observedRunningTime="2025-12-12 07:05:16.251737194 +0000 UTC m=+1003.823118463" watchObservedRunningTime="2025-12-12 07:05:16.256726006 +0000 UTC m=+1003.828107275" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.291914 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-kf2v2" podStartSLOduration=4.033089613 podStartE2EDuration="36.291894033s" podCreationTimestamp="2025-12-12 07:04:40 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.559459399 +0000 UTC m=+970.130840668" lastFinishedPulling="2025-12-12 07:05:14.818263809 +0000 UTC m=+1002.389645088" observedRunningTime="2025-12-12 07:05:16.279706576 +0000 UTC m=+1003.851087855" watchObservedRunningTime="2025-12-12 07:05:16.291894033 +0000 UTC m=+1003.863275302" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.301859 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" podStartSLOduration=5.996330089 podStartE2EDuration="35.301820585s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.321096524 +0000 UTC m=+970.892477783" lastFinishedPulling="2025-12-12 07:05:12.62658701 +0000 UTC m=+1000.197968279" observedRunningTime="2025-12-12 07:05:16.299451707 +0000 UTC m=+1003.870832976" watchObservedRunningTime="2025-12-12 07:05:16.301820585 +0000 UTC m=+1003.873201854" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.330819 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-9rhfd" podStartSLOduration=3.246567625 podStartE2EDuration="36.330799292s" podCreationTimestamp="2025-12-12 07:04:40 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.092156067 +0000 UTC m=+969.663537336" lastFinishedPulling="2025-12-12 07:05:15.176387734 +0000 UTC m=+1002.747769003" observedRunningTime="2025-12-12 07:05:16.326515507 +0000 UTC m=+1003.897896776" watchObservedRunningTime="2025-12-12 07:05:16.330799292 +0000 UTC m=+1003.902180561" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.355864 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-qwkr9" podStartSLOduration=4.013990839 podStartE2EDuration="35.355846073s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.958299467 +0000 UTC m=+970.529680736" lastFinishedPulling="2025-12-12 07:05:14.300154701 +0000 UTC m=+1001.871535970" observedRunningTime="2025-12-12 07:05:16.355628858 +0000 UTC m=+1003.927010127" watchObservedRunningTime="2025-12-12 07:05:16.355846073 +0000 UTC m=+1003.927227342" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.385248 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" podStartSLOduration=6.212636685 podStartE2EDuration="35.385200199s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.29425977 +0000 UTC m=+970.865641039" lastFinishedPulling="2025-12-12 07:05:12.466823284 +0000 UTC m=+1000.038204553" observedRunningTime="2025-12-12 07:05:16.384889051 +0000 UTC m=+1003.956270320" watchObservedRunningTime="2025-12-12 07:05:16.385200199 +0000 UTC m=+1003.956581468" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.405290 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-5zfsk" podStartSLOduration=4.668070983 podStartE2EDuration="36.405268749s" podCreationTimestamp="2025-12-12 07:04:40 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.135430168 +0000 UTC m=+969.706811437" lastFinishedPulling="2025-12-12 07:05:13.872627934 +0000 UTC m=+1001.444009203" observedRunningTime="2025-12-12 07:05:16.400565743 +0000 UTC m=+1003.971947012" watchObservedRunningTime="2025-12-12 07:05:16.405268749 +0000 UTC m=+1003.976650028" Dec 12 07:05:16 crc kubenswrapper[4867]: I1212 07:05:16.430110 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" podStartSLOduration=6.292255073 podStartE2EDuration="35.430091844s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.306170275 +0000 UTC m=+970.877551544" lastFinishedPulling="2025-12-12 07:05:12.444007046 +0000 UTC m=+1000.015388315" observedRunningTime="2025-12-12 07:05:16.429611872 +0000 UTC m=+1004.000993141" watchObservedRunningTime="2025-12-12 07:05:16.430091844 +0000 UTC m=+1004.001473113" Dec 12 07:05:17 crc kubenswrapper[4867]: I1212 07:05:17.135393 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" event={"ID":"09951d5e-0bdd-4a76-b433-880fa1bb7abd","Type":"ContainerStarted","Data":"aa3180559fd4a8f8c375a50a7931faad60d856b2d0689d0e35e0dc1062a5d854"} Dec 12 07:05:17 crc kubenswrapper[4867]: I1212 07:05:17.135746 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:05:17 crc kubenswrapper[4867]: I1212 07:05:17.140924 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" event={"ID":"9c8a7928-c0c9-4516-8e34-273afc40ccae","Type":"ContainerStarted","Data":"7291f11cf3701327bbccaeb22c53970cdc218a7e01745fc2999b2c294f65c0c1"} Dec 12 07:05:17 crc kubenswrapper[4867]: I1212 07:05:17.170739 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" podStartSLOduration=2.528575049 podStartE2EDuration="36.170718199s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:42.988198938 +0000 UTC m=+970.559580207" lastFinishedPulling="2025-12-12 07:05:16.630342078 +0000 UTC m=+1004.201723357" observedRunningTime="2025-12-12 07:05:17.166218339 +0000 UTC m=+1004.737599608" watchObservedRunningTime="2025-12-12 07:05:17.170718199 +0000 UTC m=+1004.742099478" Dec 12 07:05:17 crc kubenswrapper[4867]: I1212 07:05:17.171124 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" podStartSLOduration=2.438531492 podStartE2EDuration="36.171117319s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:04:43.072121964 +0000 UTC m=+970.643503233" lastFinishedPulling="2025-12-12 07:05:16.804707791 +0000 UTC m=+1004.376089060" observedRunningTime="2025-12-12 07:05:17.15352199 +0000 UTC m=+1004.724903259" watchObservedRunningTime="2025-12-12 07:05:17.171117319 +0000 UTC m=+1004.742498588" Dec 12 07:05:18 crc kubenswrapper[4867]: I1212 07:05:18.132869 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5b74fbd87-zjj58" Dec 12 07:05:18 crc kubenswrapper[4867]: I1212 07:05:18.150293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" event={"ID":"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6","Type":"ContainerStarted","Data":"bb162198b18b53bdb05032f14ea10826bca15b1770e422bd7db41cae21650e95"} Dec 12 07:05:18 crc kubenswrapper[4867]: I1212 07:05:18.150372 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" event={"ID":"6c74d9ef-6d91-4c62-bfe4-e4811b78adf6","Type":"ContainerStarted","Data":"cab7a99a05035855c70b90a8af44ca70908a8fdf3c86914d8658fc5480d837b3"} Dec 12 07:05:18 crc kubenswrapper[4867]: I1212 07:05:18.151596 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:05:18 crc kubenswrapper[4867]: I1212 07:05:18.187015 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" podStartSLOduration=33.716569988 podStartE2EDuration="37.187000428s" podCreationTimestamp="2025-12-12 07:04:41 +0000 UTC" firstStartedPulling="2025-12-12 07:05:14.261237722 +0000 UTC m=+1001.832618991" lastFinishedPulling="2025-12-12 07:05:17.731668162 +0000 UTC m=+1005.303049431" observedRunningTime="2025-12-12 07:05:18.185111142 +0000 UTC m=+1005.756492411" watchObservedRunningTime="2025-12-12 07:05:18.187000428 +0000 UTC m=+1005.758381697" Dec 12 07:05:19 crc kubenswrapper[4867]: I1212 07:05:19.156627 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:05:21 crc kubenswrapper[4867]: I1212 07:05:21.593399 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-62qkh" Dec 12 07:05:21 crc kubenswrapper[4867]: I1212 07:05:21.717207 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-6drsk" Dec 12 07:05:21 crc kubenswrapper[4867]: I1212 07:05:21.821196 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-c8dlk" Dec 12 07:05:22 crc kubenswrapper[4867]: I1212 07:05:22.151020 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-r24lf" Dec 12 07:05:22 crc kubenswrapper[4867]: I1212 07:05:22.172594 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vhfhv" Dec 12 07:05:22 crc kubenswrapper[4867]: I1212 07:05:22.355884 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-vnbsx" Dec 12 07:05:22 crc kubenswrapper[4867]: I1212 07:05:22.364895 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-q84bc" Dec 12 07:05:22 crc kubenswrapper[4867]: I1212 07:05:22.411149 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-4cfxw" Dec 12 07:05:23 crc kubenswrapper[4867]: I1212 07:05:23.294555 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-jqlwv" Dec 12 07:05:27 crc kubenswrapper[4867]: I1212 07:05:27.707056 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5c747c4-kzx62" Dec 12 07:05:28 crc kubenswrapper[4867]: I1212 07:05:28.988806 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:05:28 crc kubenswrapper[4867]: I1212 07:05:28.989153 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.505065 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.508337 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.510285 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rv4gh" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.513828 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.514082 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.514339 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.522369 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.522478 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26zcf\" (UniqueName: \"kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.526181 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.576025 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.577180 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.583035 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.623570 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.623989 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.624019 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmfh\" (UniqueName: \"kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.624308 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26zcf\" (UniqueName: \"kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.624386 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.624446 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.628854 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.645116 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26zcf\" (UniqueName: \"kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf\") pod \"dnsmasq-dns-84bb9d8bd9-rtc7v\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.725359 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.725441 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.725463 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmfh\" (UniqueName: \"kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.726324 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.726778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.745290 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmfh\" (UniqueName: \"kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh\") pod \"dnsmasq-dns-5f854695bc-tqbp2\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.829210 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:05:44 crc kubenswrapper[4867]: I1212 07:05:44.893736 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:05:45 crc kubenswrapper[4867]: I1212 07:05:45.401649 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:05:45 crc kubenswrapper[4867]: I1212 07:05:45.840506 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:05:45 crc kubenswrapper[4867]: W1212 07:05:45.843197 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaae8956d_1e7c_4419_83a3_a516496997fd.slice/crio-88ccbd47c3ebe860d3f9e82caf677239f9302d136944ec755b8d0d030f3bbb14 WatchSource:0}: Error finding container 88ccbd47c3ebe860d3f9e82caf677239f9302d136944ec755b8d0d030f3bbb14: Status 404 returned error can't find the container with id 88ccbd47c3ebe860d3f9e82caf677239f9302d136944ec755b8d0d030f3bbb14 Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.341509 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" event={"ID":"75b91e01-75f7-4b0b-9dc4-282572449c93","Type":"ContainerStarted","Data":"3930c984e648d86f528554b48a59272eecba910da86b7d7470aeab7a30f99605"} Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.344003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" event={"ID":"aae8956d-1e7c-4419-83a3-a516496997fd","Type":"ContainerStarted","Data":"88ccbd47c3ebe860d3f9e82caf677239f9302d136944ec755b8d0d030f3bbb14"} Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.715498 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.753142 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.754689 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.771859 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.857611 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nglxw\" (UniqueName: \"kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.857689 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.857817 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.959263 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nglxw\" (UniqueName: \"kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.959326 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.959423 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.960476 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.962070 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:46 crc kubenswrapper[4867]: I1212 07:05:46.984068 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nglxw\" (UniqueName: \"kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw\") pod \"dnsmasq-dns-c7cbb8f79-xwl9q\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.098938 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.488996 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.542165 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.545740 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.559786 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.571053 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.571265 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtx8h\" (UniqueName: \"kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.571362 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.621188 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:05:47 crc kubenswrapper[4867]: W1212 07:05:47.643387 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3d4991b_af99_4798_be05_db9e13dde565.slice/crio-ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c WatchSource:0}: Error finding container ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c: Status 404 returned error can't find the container with id ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.672568 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtx8h\" (UniqueName: \"kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.672621 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.672686 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.673835 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.675655 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.697538 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtx8h\" (UniqueName: \"kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h\") pod \"dnsmasq-dns-95f5f6995-rslfd\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.875480 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.943858 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.945587 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.947852 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.956635 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.957008 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.957159 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.958698 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.958727 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.958761 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 12 07:05:47 crc kubenswrapper[4867]: I1212 07:05:47.961494 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6hvsk" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081072 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081165 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081189 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081323 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081388 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081407 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081508 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081542 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jtzq\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081567 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.081592 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183190 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183257 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jtzq\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183278 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183298 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183360 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183415 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183439 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183503 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183540 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.183555 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.184314 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.184615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.184894 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.185305 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.186238 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.188960 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.189382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.192682 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.196801 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.200582 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jtzq\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.206339 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.217792 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.337542 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.639628 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.659547 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 07:05:48 crc kubenswrapper[4867]: I1212 07:05:48.666433 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.153168 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.165458 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.165834 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.166022 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.166297 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.166672 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.167073 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x6qvh" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.167419 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.222957 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" event={"ID":"d3d4991b-af99-4798-be05-db9e13dde565","Type":"ContainerStarted","Data":"ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c"} Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253088 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253286 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253327 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn89r\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253354 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253381 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253410 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253444 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253506 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253539 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253569 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.253595 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354402 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354460 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn89r\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354489 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354511 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354540 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354563 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354620 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354641 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354663 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354679 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.354701 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.356741 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.356802 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.357335 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.358585 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.358806 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.359053 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.373508 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.385094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.385727 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.396803 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.411780 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn89r\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.423349 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.484806 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.839838 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:05:49 crc kubenswrapper[4867]: I1212 07:05:49.967836 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.240293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" event={"ID":"87d29eb9-e24f-417c-b71f-51bbb63d3a58","Type":"ContainerStarted","Data":"2917adab8cd8fc51b4fe0bbebd42cee593c1b4c4eca865dc2f6bf169f51c60dd"} Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.570854 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.575321 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.580665 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.580810 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.582146 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lpdhb" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.583414 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.586637 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.588621 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679295 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679360 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679383 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679450 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679484 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679693 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgvv9\" (UniqueName: \"kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.679730 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.780994 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgvv9\" (UniqueName: \"kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781081 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781119 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781147 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781171 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781252 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781308 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.781520 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.782010 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.782578 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.783244 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.785094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.797179 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.797215 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.807937 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgvv9\" (UniqueName: \"kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.819939 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " pod="openstack/openstack-galera-0" Dec 12 07:05:50 crc kubenswrapper[4867]: I1212 07:05:50.892415 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.316783 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.318469 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.330389 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.330599 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.330670 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.330880 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6pzpz" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.339043 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390649 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390712 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390837 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390904 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390924 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.390947 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brbn5\" (UniqueName: \"kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.391178 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.391484 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493343 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493431 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493462 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493543 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493569 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493588 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493615 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brbn5\" (UniqueName: \"kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493659 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.493952 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.494476 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.495074 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.495333 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.495799 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.501447 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.516605 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.523096 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brbn5\" (UniqueName: \"kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.534698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.639572 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.718740 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.723767 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.726650 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-swng6" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.726845 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.727568 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.731965 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.799166 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.799261 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.799294 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.799312 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.799363 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg8ql\" (UniqueName: \"kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.903071 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.903131 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.903256 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg8ql\" (UniqueName: \"kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.903325 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.903395 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.904114 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.904328 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.907451 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.910033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:51 crc kubenswrapper[4867]: I1212 07:05:51.927726 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg8ql\" (UniqueName: \"kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql\") pod \"memcached-0\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " pod="openstack/memcached-0" Dec 12 07:05:52 crc kubenswrapper[4867]: I1212 07:05:52.063448 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.698637 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.701314 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.703467 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-z99hg" Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.706819 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.831212 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnk2g\" (UniqueName: \"kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g\") pod \"kube-state-metrics-0\" (UID: \"344aed5c-2f3f-405c-a961-c31e19468100\") " pod="openstack/kube-state-metrics-0" Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.933239 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnk2g\" (UniqueName: \"kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g\") pod \"kube-state-metrics-0\" (UID: \"344aed5c-2f3f-405c-a961-c31e19468100\") " pod="openstack/kube-state-metrics-0" Dec 12 07:05:53 crc kubenswrapper[4867]: I1212 07:05:53.953303 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnk2g\" (UniqueName: \"kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g\") pod \"kube-state-metrics-0\" (UID: \"344aed5c-2f3f-405c-a961-c31e19468100\") " pod="openstack/kube-state-metrics-0" Dec 12 07:05:54 crc kubenswrapper[4867]: I1212 07:05:54.026054 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.948360 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.949932 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.954282 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tswcb" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.954430 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.954625 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.957096 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.958786 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.969239 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:05:56 crc kubenswrapper[4867]: I1212 07:05:56.975053 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084168 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084288 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084317 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084359 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084605 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084778 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084857 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcqbl\" (UniqueName: \"kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084931 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.084990 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.085065 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.085102 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z5kv\" (UniqueName: \"kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.186999 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187051 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187092 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187110 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187134 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187166 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187185 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187208 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187240 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcqbl\" (UniqueName: \"kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187262 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187291 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187318 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.187339 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z5kv\" (UniqueName: \"kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188317 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188396 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188618 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188692 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188723 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.188868 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.189371 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.189717 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.190872 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.193284 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.196351 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.206729 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z5kv\" (UniqueName: \"kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv\") pod \"ovn-controller-ovs-9k5zx\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.209487 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcqbl\" (UniqueName: \"kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl\") pod \"ovn-controller-8njtj\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.273136 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.289743 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.292262 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerStarted","Data":"bf7d57e9df490ee931e5c05133417a009be1edebd8eca9d810067ef809aee181"} Dec 12 07:05:57 crc kubenswrapper[4867]: I1212 07:05:57.293453 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerStarted","Data":"cfbbc9e4351cbf29e8bafa83fd622c0cc562bf053e29dadb8d8188072ae0f4ec"} Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.703976 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.706771 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.725633 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.727249 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.727485 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.728037 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-955gz" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.730431 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.733839 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.895910 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896033 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8hh8\" (UniqueName: \"kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896069 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896117 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896217 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.896274 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.989123 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.989187 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.989253 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.990084 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.990154 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713" gracePeriod=600 Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997612 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997758 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8hh8\" (UniqueName: \"kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997791 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997818 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997845 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997883 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997936 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.997976 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.998369 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.998603 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:58 crc kubenswrapper[4867]: I1212 07:05:58.998862 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.003256 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.004045 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.011766 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.020015 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8hh8\" (UniqueName: \"kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.030078 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.045931 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " pod="openstack/ovsdbserver-nb-0" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.319035 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713" exitCode=0 Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.319414 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713"} Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.319462 4867 scope.go:117] "RemoveContainer" containerID="f50fc92c2a5a4121dac171c86d47138cef828299fece0f3fde41c3719947ac14" Dec 12 07:05:59 crc kubenswrapper[4867]: I1212 07:05:59.328184 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.564370 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.565940 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.568415 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.568545 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.568670 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jl59v" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.575616 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.578115 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.730819 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.730877 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.730899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n5tf\" (UniqueName: \"kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.730925 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.731022 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.731048 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.731069 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.731096 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833000 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833065 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833092 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833128 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833196 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833246 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833269 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n5tf\" (UniqueName: \"kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833294 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833504 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.833818 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.834345 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.837843 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.839299 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.841432 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.845265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.858710 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n5tf\" (UniqueName: \"kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.866935 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:00 crc kubenswrapper[4867]: I1212 07:06:00.889287 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.703063 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.703799 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lrmfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5f854695bc-tqbp2_openstack(75b91e01-75f7-4b0b-9dc4-282572449c93): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.705969 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" podUID="75b91e01-75f7-4b0b-9dc4-282572449c93" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.762671 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.762814 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-26zcf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84bb9d8bd9-rtc7v_openstack(aae8956d-1e7c-4419-83a3-a516496997fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:06:07 crc kubenswrapper[4867]: E1212 07:06:07.764211 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" podUID="aae8956d-1e7c-4419-83a3-a516496997fd" Dec 12 07:06:08 crc kubenswrapper[4867]: I1212 07:06:08.188170 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:06:09 crc kubenswrapper[4867]: W1212 07:06:09.150848 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ce609f7_1106_4d01_9a6b_d2099112ce33.slice/crio-66da9a2b24b25a23ddff48c1b4c394ad7a15395a4892c18dbf7f7609a63457e3 WatchSource:0}: Error finding container 66da9a2b24b25a23ddff48c1b4c394ad7a15395a4892c18dbf7f7609a63457e3: Status 404 returned error can't find the container with id 66da9a2b24b25a23ddff48c1b4c394ad7a15395a4892c18dbf7f7609a63457e3 Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.268479 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.287624 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc\") pod \"75b91e01-75f7-4b0b-9dc4-282572449c93\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.287706 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrmfh\" (UniqueName: \"kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh\") pod \"75b91e01-75f7-4b0b-9dc4-282572449c93\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.287745 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config\") pod \"75b91e01-75f7-4b0b-9dc4-282572449c93\" (UID: \"75b91e01-75f7-4b0b-9dc4-282572449c93\") " Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.288287 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75b91e01-75f7-4b0b-9dc4-282572449c93" (UID: "75b91e01-75f7-4b0b-9dc4-282572449c93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.288640 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config" (OuterVolumeSpecName: "config") pod "75b91e01-75f7-4b0b-9dc4-282572449c93" (UID: "75b91e01-75f7-4b0b-9dc4-282572449c93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.292055 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh" (OuterVolumeSpecName: "kube-api-access-lrmfh") pod "75b91e01-75f7-4b0b-9dc4-282572449c93" (UID: "75b91e01-75f7-4b0b-9dc4-282572449c93"). InnerVolumeSpecName "kube-api-access-lrmfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.320528 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.388900 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config\") pod \"aae8956d-1e7c-4419-83a3-a516496997fd\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.389293 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26zcf\" (UniqueName: \"kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf\") pod \"aae8956d-1e7c-4419-83a3-a516496997fd\" (UID: \"aae8956d-1e7c-4419-83a3-a516496997fd\") " Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.389786 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config" (OuterVolumeSpecName: "config") pod "aae8956d-1e7c-4419-83a3-a516496997fd" (UID: "aae8956d-1e7c-4419-83a3-a516496997fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.390060 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae8956d-1e7c-4419-83a3-a516496997fd-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.390086 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.390098 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrmfh\" (UniqueName: \"kubernetes.io/projected/75b91e01-75f7-4b0b-9dc4-282572449c93-kube-api-access-lrmfh\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.390111 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75b91e01-75f7-4b0b-9dc4-282572449c93-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.395435 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf" (OuterVolumeSpecName: "kube-api-access-26zcf") pod "aae8956d-1e7c-4419-83a3-a516496997fd" (UID: "aae8956d-1e7c-4419-83a3-a516496997fd"). InnerVolumeSpecName "kube-api-access-26zcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.407116 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" event={"ID":"75b91e01-75f7-4b0b-9dc4-282572449c93","Type":"ContainerDied","Data":"3930c984e648d86f528554b48a59272eecba910da86b7d7470aeab7a30f99605"} Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.407207 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-tqbp2" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.413571 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerStarted","Data":"66da9a2b24b25a23ddff48c1b4c394ad7a15395a4892c18dbf7f7609a63457e3"} Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.428408 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" event={"ID":"aae8956d-1e7c-4419-83a3-a516496997fd","Type":"ContainerDied","Data":"88ccbd47c3ebe860d3f9e82caf677239f9302d136944ec755b8d0d030f3bbb14"} Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.428714 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-rtc7v" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.493693 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26zcf\" (UniqueName: \"kubernetes.io/projected/aae8956d-1e7c-4419-83a3-a516496997fd-kube-api-access-26zcf\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.601059 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.624488 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-tqbp2"] Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.642614 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.664698 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.689208 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-rtc7v"] Dec 12 07:06:09 crc kubenswrapper[4867]: I1212 07:06:09.952382 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.062509 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:06:10 crc kubenswrapper[4867]: W1212 07:06:10.115927 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e39305a_805b_4e9f_ac88_e85e8007409f.slice/crio-055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6 WatchSource:0}: Error finding container 055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6: Status 404 returned error can't find the container with id 055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6 Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.118679 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.128763 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.193967 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:06:10 crc kubenswrapper[4867]: W1212 07:06:10.197706 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04569a7e_a52f_4c83_8d2d_e2d6c1ac6d6f.slice/crio-5807753ddb1090827c604ebc45d5015de14ea2b2c68ee8978aefc0d37f32810a WatchSource:0}: Error finding container 5807753ddb1090827c604ebc45d5015de14ea2b2c68ee8978aefc0d37f32810a: Status 404 returned error can't find the container with id 5807753ddb1090827c604ebc45d5015de14ea2b2c68ee8978aefc0d37f32810a Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.439169 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerStarted","Data":"7a3352d8bd3e4131024a0604a55fee9457316c1ddd8d1eeff2808f0dd1f89a30"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.440700 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"45fe5995-c9da-4b48-bef1-20cc51f101f2","Type":"ContainerStarted","Data":"3105b7f04980db6866b53c011e49ea9c426dd90bf4770bf93d92474332028e8b"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.442361 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerStarted","Data":"5807753ddb1090827c604ebc45d5015de14ea2b2c68ee8978aefc0d37f32810a"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.444599 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerStarted","Data":"055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.446358 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj" event={"ID":"9a11a089-0048-4c70-94e5-9ae18eadfed5","Type":"ContainerStarted","Data":"aebb55f232f65de131c725eada57789830eee1d462d92db26d07c3d12f0c96f3"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.447715 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"344aed5c-2f3f-405c-a961-c31e19468100","Type":"ContainerStarted","Data":"3e166c4ab0b1a8eec98fe03e56315537b792052fef2ad7865a92f1bd078dc22e"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.451071 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.454179 4867 generic.go:334] "Generic (PLEG): container finished" podID="d3d4991b-af99-4798-be05-db9e13dde565" containerID="7ccd1899441c63af5cdb1d2a0b868ca031e3b4606a91aa7f75fd4116d1d34445" exitCode=0 Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.454293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" event={"ID":"d3d4991b-af99-4798-be05-db9e13dde565","Type":"ContainerDied","Data":"7ccd1899441c63af5cdb1d2a0b868ca031e3b4606a91aa7f75fd4116d1d34445"} Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.458238 4867 generic.go:334] "Generic (PLEG): container finished" podID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerID="f17004fdeeb584465b109ce2494ff8b3ebe6be7759298ba2e3617b73e515a789" exitCode=0 Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.458313 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" event={"ID":"87d29eb9-e24f-417c-b71f-51bbb63d3a58","Type":"ContainerDied","Data":"f17004fdeeb584465b109ce2494ff8b3ebe6be7759298ba2e3617b73e515a789"} Dec 12 07:06:10 crc kubenswrapper[4867]: E1212 07:06:10.828660 4867 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 12 07:06:10 crc kubenswrapper[4867]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/d3d4991b-af99-4798-be05-db9e13dde565/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 12 07:06:10 crc kubenswrapper[4867]: > podSandboxID="ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c" Dec 12 07:06:10 crc kubenswrapper[4867]: E1212 07:06:10.829030 4867 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 12 07:06:10 crc kubenswrapper[4867]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nglxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-c7cbb8f79-xwl9q_openstack(d3d4991b-af99-4798-be05-db9e13dde565): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/d3d4991b-af99-4798-be05-db9e13dde565/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 12 07:06:10 crc kubenswrapper[4867]: > logger="UnhandledError" Dec 12 07:06:10 crc kubenswrapper[4867]: E1212 07:06:10.830187 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/d3d4991b-af99-4798-be05-db9e13dde565/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" podUID="d3d4991b-af99-4798-be05-db9e13dde565" Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.849788 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b91e01-75f7-4b0b-9dc4-282572449c93" path="/var/lib/kubelet/pods/75b91e01-75f7-4b0b-9dc4-282572449c93/volumes" Dec 12 07:06:10 crc kubenswrapper[4867]: I1212 07:06:10.850278 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae8956d-1e7c-4419-83a3-a516496997fd" path="/var/lib/kubelet/pods/aae8956d-1e7c-4419-83a3-a516496997fd/volumes" Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.077513 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.469537 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerStarted","Data":"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020"} Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.471981 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerStarted","Data":"0f947d9d4fe9c6b4d937ad0d0687b9d1db4c6172ed3a4aac66489775077c3e68"} Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.493192 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerStarted","Data":"22e955d75ce13bcaae31e349f84ce73ac23e943d07cc4217976ebd4b9e74c392"} Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.504338 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" event={"ID":"87d29eb9-e24f-417c-b71f-51bbb63d3a58","Type":"ContainerStarted","Data":"20a113b51f1390fced95e76529377d456fd6c7eb03779413cec0fa70965c1ac2"} Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.504551 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:06:11 crc kubenswrapper[4867]: I1212 07:06:11.568762 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" podStartSLOduration=4.545484369 podStartE2EDuration="24.568746955s" podCreationTimestamp="2025-12-12 07:05:47 +0000 UTC" firstStartedPulling="2025-12-12 07:05:49.222107679 +0000 UTC m=+1036.793488948" lastFinishedPulling="2025-12-12 07:06:09.245370275 +0000 UTC m=+1056.816751534" observedRunningTime="2025-12-12 07:06:11.564016151 +0000 UTC m=+1059.135397450" watchObservedRunningTime="2025-12-12 07:06:11.568746955 +0000 UTC m=+1059.140128224" Dec 12 07:06:17 crc kubenswrapper[4867]: I1212 07:06:17.877380 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:06:17 crc kubenswrapper[4867]: I1212 07:06:17.941727 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.593576 4867 generic.go:334] "Generic (PLEG): container finished" podID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerID="f5d8542ed11ac48acd676d04434b7997709cc56b7d071495b3ebe50031d5d89c" exitCode=0 Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.593640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerDied","Data":"f5d8542ed11ac48acd676d04434b7997709cc56b7d071495b3ebe50031d5d89c"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.596920 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" event={"ID":"d3d4991b-af99-4798-be05-db9e13dde565","Type":"ContainerStarted","Data":"9b0c066ec3779c7a23f46525752cba5b87ca94807dcf8f9c2e4a83db8e72a719"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.597048 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="dnsmasq-dns" containerID="cri-o://9b0c066ec3779c7a23f46525752cba5b87ca94807dcf8f9c2e4a83db8e72a719" gracePeriod=10 Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.597090 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.599241 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerStarted","Data":"e8768b0eadae644cd43b37f9b16a6da9182b19bd8b06c8288046607848e24de1"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.602519 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerStarted","Data":"a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.605710 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerStarted","Data":"d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.609081 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj" event={"ID":"9a11a089-0048-4c70-94e5-9ae18eadfed5","Type":"ContainerStarted","Data":"43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.609248 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8njtj" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.611601 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"45fe5995-c9da-4b48-bef1-20cc51f101f2","Type":"ContainerStarted","Data":"b4b21d46b7b68ccba3a77319862cb744cd50115fe97c79196a5c300c08368180"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.612406 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.614132 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerStarted","Data":"e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.624018 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"344aed5c-2f3f-405c-a961-c31e19468100","Type":"ContainerStarted","Data":"f31de2a58bb8c0627835f82641e74392d8eaa399cc5779ec9d25f0762802a2c6"} Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.624861 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.652031 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8njtj" podStartSLOduration=15.325750726 podStartE2EDuration="25.652014763s" podCreationTimestamp="2025-12-12 07:05:56 +0000 UTC" firstStartedPulling="2025-12-12 07:06:10.133750674 +0000 UTC m=+1057.705131943" lastFinishedPulling="2025-12-12 07:06:20.460014711 +0000 UTC m=+1068.031395980" observedRunningTime="2025-12-12 07:06:21.651023758 +0000 UTC m=+1069.222405027" watchObservedRunningTime="2025-12-12 07:06:21.652014763 +0000 UTC m=+1069.223396022" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.704184 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.191954996 podStartE2EDuration="30.704145718s" podCreationTimestamp="2025-12-12 07:05:51 +0000 UTC" firstStartedPulling="2025-12-12 07:06:09.969325932 +0000 UTC m=+1057.540707201" lastFinishedPulling="2025-12-12 07:06:20.481516654 +0000 UTC m=+1068.052897923" observedRunningTime="2025-12-12 07:06:21.702353435 +0000 UTC m=+1069.273734724" watchObservedRunningTime="2025-12-12 07:06:21.704145718 +0000 UTC m=+1069.275526987" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.728439 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=18.396722617 podStartE2EDuration="28.728404347s" podCreationTimestamp="2025-12-12 07:05:53 +0000 UTC" firstStartedPulling="2025-12-12 07:06:10.131303784 +0000 UTC m=+1057.702685053" lastFinishedPulling="2025-12-12 07:06:20.462985514 +0000 UTC m=+1068.034366783" observedRunningTime="2025-12-12 07:06:21.688939889 +0000 UTC m=+1069.260321158" watchObservedRunningTime="2025-12-12 07:06:21.728404347 +0000 UTC m=+1069.299785636" Dec 12 07:06:21 crc kubenswrapper[4867]: I1212 07:06:21.767220 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" podStartSLOduration=14.137459964 podStartE2EDuration="35.767205019s" podCreationTimestamp="2025-12-12 07:05:46 +0000 UTC" firstStartedPulling="2025-12-12 07:05:47.656366689 +0000 UTC m=+1035.227747958" lastFinishedPulling="2025-12-12 07:06:09.286111744 +0000 UTC m=+1056.857493013" observedRunningTime="2025-12-12 07:06:21.764717279 +0000 UTC m=+1069.336098558" watchObservedRunningTime="2025-12-12 07:06:21.767205019 +0000 UTC m=+1069.338586288" Dec 12 07:06:23 crc kubenswrapper[4867]: I1212 07:06:23.645949 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerStarted","Data":"fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11"} Dec 12 07:06:23 crc kubenswrapper[4867]: I1212 07:06:23.646597 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerStarted","Data":"969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4"} Dec 12 07:06:23 crc kubenswrapper[4867]: I1212 07:06:23.650259 4867 generic.go:334] "Generic (PLEG): container finished" podID="d3d4991b-af99-4798-be05-db9e13dde565" containerID="9b0c066ec3779c7a23f46525752cba5b87ca94807dcf8f9c2e4a83db8e72a719" exitCode=0 Dec 12 07:06:23 crc kubenswrapper[4867]: I1212 07:06:23.650315 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" event={"ID":"d3d4991b-af99-4798-be05-db9e13dde565","Type":"ContainerDied","Data":"9b0c066ec3779c7a23f46525752cba5b87ca94807dcf8f9c2e4a83db8e72a719"} Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.461523 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.550541 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc\") pod \"d3d4991b-af99-4798-be05-db9e13dde565\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.550727 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config\") pod \"d3d4991b-af99-4798-be05-db9e13dde565\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.550763 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nglxw\" (UniqueName: \"kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw\") pod \"d3d4991b-af99-4798-be05-db9e13dde565\" (UID: \"d3d4991b-af99-4798-be05-db9e13dde565\") " Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.570475 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw" (OuterVolumeSpecName: "kube-api-access-nglxw") pod "d3d4991b-af99-4798-be05-db9e13dde565" (UID: "d3d4991b-af99-4798-be05-db9e13dde565"). InnerVolumeSpecName "kube-api-access-nglxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.591851 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3d4991b-af99-4798-be05-db9e13dde565" (UID: "d3d4991b-af99-4798-be05-db9e13dde565"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.605522 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config" (OuterVolumeSpecName: "config") pod "d3d4991b-af99-4798-be05-db9e13dde565" (UID: "d3d4991b-af99-4798-be05-db9e13dde565"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.653449 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.653479 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nglxw\" (UniqueName: \"kubernetes.io/projected/d3d4991b-af99-4798-be05-db9e13dde565-kube-api-access-nglxw\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.653492 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3d4991b-af99-4798-be05-db9e13dde565-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.661443 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" event={"ID":"d3d4991b-af99-4798-be05-db9e13dde565","Type":"ContainerDied","Data":"ecdf1238761376a80ead4c359cf71fe242b71bc523abadc6995376d00528b56c"} Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.661534 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.661601 4867 scope.go:117] "RemoveContainer" containerID="9b0c066ec3779c7a23f46525752cba5b87ca94807dcf8f9c2e4a83db8e72a719" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.661624 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.661614 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7cbb8f79-xwl9q" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.688120 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-9k5zx" podStartSLOduration=19.600495526 podStartE2EDuration="28.688098028s" podCreationTimestamp="2025-12-12 07:05:56 +0000 UTC" firstStartedPulling="2025-12-12 07:06:11.383195801 +0000 UTC m=+1058.954577070" lastFinishedPulling="2025-12-12 07:06:20.470798303 +0000 UTC m=+1068.042179572" observedRunningTime="2025-12-12 07:06:24.680511884 +0000 UTC m=+1072.251893173" watchObservedRunningTime="2025-12-12 07:06:24.688098028 +0000 UTC m=+1072.259479297" Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.697911 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.703675 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c7cbb8f79-xwl9q"] Dec 12 07:06:24 crc kubenswrapper[4867]: I1212 07:06:24.852204 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3d4991b-af99-4798-be05-db9e13dde565" path="/var/lib/kubelet/pods/d3d4991b-af99-4798-be05-db9e13dde565/volumes" Dec 12 07:06:25 crc kubenswrapper[4867]: I1212 07:06:25.956539 4867 scope.go:117] "RemoveContainer" containerID="7ccd1899441c63af5cdb1d2a0b868ca031e3b4606a91aa7f75fd4116d1d34445" Dec 12 07:06:26 crc kubenswrapper[4867]: I1212 07:06:26.673993 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerStarted","Data":"ef733b2c26cb5cf662dfa06f03ae9d9a2ad1b9bfc775896fa039f832b4e8f2b0"} Dec 12 07:06:26 crc kubenswrapper[4867]: I1212 07:06:26.675973 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerStarted","Data":"48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098"} Dec 12 07:06:26 crc kubenswrapper[4867]: I1212 07:06:26.697501 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.844213994 podStartE2EDuration="27.697479695s" podCreationTimestamp="2025-12-12 07:05:59 +0000 UTC" firstStartedPulling="2025-12-12 07:06:10.203624021 +0000 UTC m=+1057.775005290" lastFinishedPulling="2025-12-12 07:06:26.056889722 +0000 UTC m=+1073.628270991" observedRunningTime="2025-12-12 07:06:26.692680098 +0000 UTC m=+1074.264061367" watchObservedRunningTime="2025-12-12 07:06:26.697479695 +0000 UTC m=+1074.268860964" Dec 12 07:06:26 crc kubenswrapper[4867]: I1212 07:06:26.717799 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.794573179 podStartE2EDuration="29.717782868s" podCreationTimestamp="2025-12-12 07:05:57 +0000 UTC" firstStartedPulling="2025-12-12 07:06:10.120715788 +0000 UTC m=+1057.692097057" lastFinishedPulling="2025-12-12 07:06:26.043925477 +0000 UTC m=+1073.615306746" observedRunningTime="2025-12-12 07:06:26.713378951 +0000 UTC m=+1074.284760230" watchObservedRunningTime="2025-12-12 07:06:26.717782868 +0000 UTC m=+1074.289164137" Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.064663 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.685574 4867 generic.go:334] "Generic (PLEG): container finished" podID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerID="e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e" exitCode=0 Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.685650 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerDied","Data":"e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e"} Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.688791 4867 generic.go:334] "Generic (PLEG): container finished" podID="5e79e315-4744-4806-96fc-284d97fa88e0" containerID="e8768b0eadae644cd43b37f9b16a6da9182b19bd8b06c8288046607848e24de1" exitCode=0 Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.688970 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerDied","Data":"e8768b0eadae644cd43b37f9b16a6da9182b19bd8b06c8288046607848e24de1"} Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.890432 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:27 crc kubenswrapper[4867]: I1212 07:06:27.936389 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.697981 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerStarted","Data":"456526542b9c5797b244ddfdcbdceb1b0f47fb01b1e916fbb187836108af239b"} Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.701431 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerStarted","Data":"49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e"} Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.701706 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.720699 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=28.846811505 podStartE2EDuration="39.720684457s" podCreationTimestamp="2025-12-12 07:05:49 +0000 UTC" firstStartedPulling="2025-12-12 07:06:09.626086069 +0000 UTC m=+1057.197467338" lastFinishedPulling="2025-12-12 07:06:20.499959021 +0000 UTC m=+1068.071340290" observedRunningTime="2025-12-12 07:06:28.717305235 +0000 UTC m=+1076.288686504" watchObservedRunningTime="2025-12-12 07:06:28.720684457 +0000 UTC m=+1076.292065726" Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.745437 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.485730837 podStartE2EDuration="38.745351826s" podCreationTimestamp="2025-12-12 07:05:50 +0000 UTC" firstStartedPulling="2025-12-12 07:06:09.175844467 +0000 UTC m=+1056.747225736" lastFinishedPulling="2025-12-12 07:06:20.435465456 +0000 UTC m=+1068.006846725" observedRunningTime="2025-12-12 07:06:28.743882021 +0000 UTC m=+1076.315263290" watchObservedRunningTime="2025-12-12 07:06:28.745351826 +0000 UTC m=+1076.316733095" Dec 12 07:06:28 crc kubenswrapper[4867]: I1212 07:06:28.745470 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.009689 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:06:29 crc kubenswrapper[4867]: E1212 07:06:29.010161 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="dnsmasq-dns" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.010186 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="dnsmasq-dns" Dec 12 07:06:29 crc kubenswrapper[4867]: E1212 07:06:29.010212 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="init" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.010254 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="init" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.010467 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3d4991b-af99-4798-be05-db9e13dde565" containerName="dnsmasq-dns" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.024769 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.024901 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.028644 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.078510 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.081884 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.085424 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.097554 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140368 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140451 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd9wz\" (UniqueName: \"kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140502 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140543 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140604 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140645 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140662 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mplcb\" (UniqueName: \"kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140741 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140756 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.140799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242762 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd9wz\" (UniqueName: \"kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242830 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242886 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242940 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.242992 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mplcb\" (UniqueName: \"kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.243037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.243059 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.243094 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.243139 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.244127 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.244936 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.245043 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.245300 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.245823 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.245967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.251407 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.251600 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.260268 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mplcb\" (UniqueName: \"kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb\") pod \"dnsmasq-dns-5b79764b65-79vhd\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.262718 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd9wz\" (UniqueName: \"kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz\") pod \"ovn-controller-metrics-92rzj\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.329179 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.329505 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.354181 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.383118 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.419061 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.436567 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.450715 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.452096 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.462154 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.468016 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.551300 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.551349 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s2sz\" (UniqueName: \"kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.551387 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.551437 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.551499 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.652724 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.652813 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.652909 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.652937 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s2sz\" (UniqueName: \"kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.652966 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.654107 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.654772 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.654827 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.655458 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.678241 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s2sz\" (UniqueName: \"kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz\") pod \"dnsmasq-dns-586b989cdc-gf5v8\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.769460 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.819721 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.970359 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.972063 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.979189 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.980192 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.980609 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.981411 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.981587 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-m2lbq" Dec 12 07:06:29 crc kubenswrapper[4867]: W1212 07:06:29.986454 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3346b14_1b14_4a53_953b_5c6a2d6ff724.slice/crio-2ba15676a9e5cc3ee0a0ef345cabf64fa0d0439855f47e16af7566abd75d3212 WatchSource:0}: Error finding container 2ba15676a9e5cc3ee0a0ef345cabf64fa0d0439855f47e16af7566abd75d3212: Status 404 returned error can't find the container with id 2ba15676a9e5cc3ee0a0ef345cabf64fa0d0439855f47e16af7566abd75d3212 Dec 12 07:06:29 crc kubenswrapper[4867]: I1212 07:06:29.993674 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.049941 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:06:30 crc kubenswrapper[4867]: W1212 07:06:30.054997 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc087ab39_7d5b_42b3_a1f7_5f3703e1d6ba.slice/crio-416dd3385a5b82f1c0ed7903b247071a9ab944f39f0054abc1ac5eb8c8849739 WatchSource:0}: Error finding container 416dd3385a5b82f1c0ed7903b247071a9ab944f39f0054abc1ac5eb8c8849739: Status 404 returned error can't find the container with id 416dd3385a5b82f1c0ed7903b247071a9ab944f39f0054abc1ac5eb8c8849739 Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060790 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060831 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060874 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65crj\" (UniqueName: \"kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060906 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060937 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060954 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.060973 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.171992 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65crj\" (UniqueName: \"kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172108 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172197 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172249 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172292 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.172512 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.174722 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.175760 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.181099 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.182041 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.182338 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.184425 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.198575 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65crj\" (UniqueName: \"kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj\") pod \"ovn-northd-0\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.358703 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.368072 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 07:06:30 crc kubenswrapper[4867]: W1212 07:06:30.374764 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6019939f_3efa_47ee_9570_b9c14cea2ecc.slice/crio-fcb7c5fe636f87cb55a5d21f5d10236ac83ff9f326e0cd394c459fa7185d6e0c WatchSource:0}: Error finding container fcb7c5fe636f87cb55a5d21f5d10236ac83ff9f326e0cd394c459fa7185d6e0c: Status 404 returned error can't find the container with id fcb7c5fe636f87cb55a5d21f5d10236ac83ff9f326e0cd394c459fa7185d6e0c Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.727588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" event={"ID":"6019939f-3efa-47ee-9570-b9c14cea2ecc","Type":"ContainerStarted","Data":"fcb7c5fe636f87cb55a5d21f5d10236ac83ff9f326e0cd394c459fa7185d6e0c"} Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.732014 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-92rzj" event={"ID":"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba","Type":"ContainerStarted","Data":"416dd3385a5b82f1c0ed7903b247071a9ab944f39f0054abc1ac5eb8c8849739"} Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.735067 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" event={"ID":"f3346b14-1b14-4a53-953b-5c6a2d6ff724","Type":"ContainerStarted","Data":"2ba15676a9e5cc3ee0a0ef345cabf64fa0d0439855f47e16af7566abd75d3212"} Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.893986 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.894036 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 12 07:06:30 crc kubenswrapper[4867]: I1212 07:06:30.984388 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:06:31 crc kubenswrapper[4867]: W1212 07:06:31.049436 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00269882_fe60_4b26_8d5c_1b34b014b191.slice/crio-f85d8ccfd37c9b440bd05e12368df61b028435fabdd7c690acedf5b5bcbf9501 WatchSource:0}: Error finding container f85d8ccfd37c9b440bd05e12368df61b028435fabdd7c690acedf5b5bcbf9501: Status 404 returned error can't find the container with id f85d8ccfd37c9b440bd05e12368df61b028435fabdd7c690acedf5b5bcbf9501 Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.640444 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.640799 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.750340 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerStarted","Data":"f85d8ccfd37c9b440bd05e12368df61b028435fabdd7c690acedf5b5bcbf9501"} Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.754511 4867 generic.go:334] "Generic (PLEG): container finished" podID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" containerID="5e8d3e1ad21e90dff072b7e26f1cfa64c50f696c853ade67ef9ac2f564b7ee09" exitCode=0 Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.754572 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" event={"ID":"f3346b14-1b14-4a53-953b-5c6a2d6ff724","Type":"ContainerDied","Data":"5e8d3e1ad21e90dff072b7e26f1cfa64c50f696c853ade67ef9ac2f564b7ee09"} Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.757837 4867 generic.go:334] "Generic (PLEG): container finished" podID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerID="ec335c45dc9da9858d39e20d3a99708169d7880067e0cfa31406c514adf32302" exitCode=0 Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.757964 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" event={"ID":"6019939f-3efa-47ee-9570-b9c14cea2ecc","Type":"ContainerDied","Data":"ec335c45dc9da9858d39e20d3a99708169d7880067e0cfa31406c514adf32302"} Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.761620 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-92rzj" event={"ID":"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba","Type":"ContainerStarted","Data":"c4b46d15a056004b20724580a9403bb36eec9ea66794fe8c2015a7db598e5ae8"} Dec 12 07:06:31 crc kubenswrapper[4867]: I1212 07:06:31.827271 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-92rzj" podStartSLOduration=2.8272478640000003 podStartE2EDuration="2.827247864s" podCreationTimestamp="2025-12-12 07:06:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:31.826126386 +0000 UTC m=+1079.397507655" watchObservedRunningTime="2025-12-12 07:06:31.827247864 +0000 UTC m=+1079.398629133" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.139460 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.249816 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mplcb\" (UniqueName: \"kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb\") pod \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.249971 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc\") pod \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.250136 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config\") pod \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.250175 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb\") pod \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\" (UID: \"f3346b14-1b14-4a53-953b-5c6a2d6ff724\") " Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.257265 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb" (OuterVolumeSpecName: "kube-api-access-mplcb") pod "f3346b14-1b14-4a53-953b-5c6a2d6ff724" (UID: "f3346b14-1b14-4a53-953b-5c6a2d6ff724"). InnerVolumeSpecName "kube-api-access-mplcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.285829 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f3346b14-1b14-4a53-953b-5c6a2d6ff724" (UID: "f3346b14-1b14-4a53-953b-5c6a2d6ff724"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.289866 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config" (OuterVolumeSpecName: "config") pod "f3346b14-1b14-4a53-953b-5c6a2d6ff724" (UID: "f3346b14-1b14-4a53-953b-5c6a2d6ff724"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.290292 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f3346b14-1b14-4a53-953b-5c6a2d6ff724" (UID: "f3346b14-1b14-4a53-953b-5c6a2d6ff724"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.358917 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.359307 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mplcb\" (UniqueName: \"kubernetes.io/projected/f3346b14-1b14-4a53-953b-5c6a2d6ff724-kube-api-access-mplcb\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.359324 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.359337 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3346b14-1b14-4a53-953b-5c6a2d6ff724-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.770037 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" event={"ID":"6019939f-3efa-47ee-9570-b9c14cea2ecc","Type":"ContainerStarted","Data":"e7af4d5bd75e5a57bd2b4d0a42ed21234adc4add57ef546e84d2fd53187091fc"} Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.770164 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.776131 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.778882 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" event={"ID":"f3346b14-1b14-4a53-953b-5c6a2d6ff724","Type":"ContainerDied","Data":"2ba15676a9e5cc3ee0a0ef345cabf64fa0d0439855f47e16af7566abd75d3212"} Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.778943 4867 scope.go:117] "RemoveContainer" containerID="5e8d3e1ad21e90dff072b7e26f1cfa64c50f696c853ade67ef9ac2f564b7ee09" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.804614 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" podStartSLOduration=3.804584933 podStartE2EDuration="3.804584933s" podCreationTimestamp="2025-12-12 07:06:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:32.798373623 +0000 UTC m=+1080.369754892" watchObservedRunningTime="2025-12-12 07:06:32.804584933 +0000 UTC m=+1080.375966202" Dec 12 07:06:32 crc kubenswrapper[4867]: I1212 07:06:32.987975 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 12 07:06:33 crc kubenswrapper[4867]: I1212 07:06:33.095508 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 12 07:06:33 crc kubenswrapper[4867]: I1212 07:06:33.784383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerStarted","Data":"0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad"} Dec 12 07:06:33 crc kubenswrapper[4867]: I1212 07:06:33.784429 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerStarted","Data":"7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844"} Dec 12 07:06:33 crc kubenswrapper[4867]: I1212 07:06:33.805991 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.259268392 podStartE2EDuration="4.805967946s" podCreationTimestamp="2025-12-12 07:06:29 +0000 UTC" firstStartedPulling="2025-12-12 07:06:31.051586321 +0000 UTC m=+1078.622967590" lastFinishedPulling="2025-12-12 07:06:32.598285875 +0000 UTC m=+1080.169667144" observedRunningTime="2025-12-12 07:06:33.800792361 +0000 UTC m=+1081.372173660" watchObservedRunningTime="2025-12-12 07:06:33.805967946 +0000 UTC m=+1081.377349215" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.003886 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.033817 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:06:34 crc kubenswrapper[4867]: E1212 07:06:34.034165 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" containerName="init" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.034181 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" containerName="init" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.034383 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" containerName="init" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.035156 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.035457 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.056763 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.102206 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbc5\" (UniqueName: \"kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.102378 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.102457 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.102501 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.102528 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204000 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204076 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204112 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204132 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204153 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbc5\" (UniqueName: \"kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.204982 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.205069 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.205109 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.205177 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.248084 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbc5\" (UniqueName: \"kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5\") pod \"dnsmasq-dns-67fdf7998c-rlltl\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.352147 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.790696 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="dnsmasq-dns" containerID="cri-o://e7af4d5bd75e5a57bd2b4d0a42ed21234adc4add57ef546e84d2fd53187091fc" gracePeriod=10 Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.791923 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 12 07:06:34 crc kubenswrapper[4867]: I1212 07:06:34.865273 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:06:34 crc kubenswrapper[4867]: W1212 07:06:34.869372 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0f75dd4_02a1_473c_a06f_c8477baa2d90.slice/crio-651ea1cdb6a0fe7d193b7083c2f24a8ef8f5dde9eb84a51e11a08099ba9f3f8d WatchSource:0}: Error finding container 651ea1cdb6a0fe7d193b7083c2f24a8ef8f5dde9eb84a51e11a08099ba9f3f8d: Status 404 returned error can't find the container with id 651ea1cdb6a0fe7d193b7083c2f24a8ef8f5dde9eb84a51e11a08099ba9f3f8d Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.162807 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.182559 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.185520 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.185818 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-p498d" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.186352 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.187707 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.201722 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.220888 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.220978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.221021 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.221075 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbqfq\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.221121 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.324493 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.324553 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.324595 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbqfq\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.324640 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.324669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.325148 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.325775 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.325980 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.326294 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.326410 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.326887 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:06:35.826862704 +0000 UTC m=+1083.398244043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.358686 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.362012 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbqfq\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.734052 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.802010 4867 generic.go:334] "Generic (PLEG): container finished" podID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerID="e7af4d5bd75e5a57bd2b4d0a42ed21234adc4add57ef546e84d2fd53187091fc" exitCode=0 Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.802093 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" event={"ID":"6019939f-3efa-47ee-9570-b9c14cea2ecc","Type":"ContainerDied","Data":"e7af4d5bd75e5a57bd2b4d0a42ed21234adc4add57ef546e84d2fd53187091fc"} Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.804730 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" event={"ID":"b0f75dd4-02a1-473c-a06f-c8477baa2d90","Type":"ContainerStarted","Data":"651ea1cdb6a0fe7d193b7083c2f24a8ef8f5dde9eb84a51e11a08099ba9f3f8d"} Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.820796 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 12 07:06:35 crc kubenswrapper[4867]: I1212 07:06:35.833112 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.833853 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.833893 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:35 crc kubenswrapper[4867]: E1212 07:06:35.833942 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:06:36.833921485 +0000 UTC m=+1084.405302844 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:36 crc kubenswrapper[4867]: I1212 07:06:36.851737 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:36 crc kubenswrapper[4867]: E1212 07:06:36.851944 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:36 crc kubenswrapper[4867]: E1212 07:06:36.852154 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:36 crc kubenswrapper[4867]: E1212 07:06:36.852208 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:06:38.852190078 +0000 UTC m=+1086.423571347 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.312505 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a6d9-account-create-update-s9hgd"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.313864 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.316817 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.321347 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a6d9-account-create-update-s9hgd"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.376930 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-znq7k"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.378002 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.388404 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-znq7k"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.470339 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.470439 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5b9j\" (UniqueName: \"kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.545959 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.573962 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.574020 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.574039 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp6lv\" (UniqueName: \"kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.574075 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5b9j\" (UniqueName: \"kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.575771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.591602 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5b9j\" (UniqueName: \"kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j\") pod \"glance-a6d9-account-create-update-s9hgd\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.639968 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s2sz\" (UniqueName: \"kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz\") pod \"6019939f-3efa-47ee-9570-b9c14cea2ecc\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config\") pod \"6019939f-3efa-47ee-9570-b9c14cea2ecc\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675470 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc\") pod \"6019939f-3efa-47ee-9570-b9c14cea2ecc\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675574 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb\") pod \"6019939f-3efa-47ee-9570-b9c14cea2ecc\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675599 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb\") pod \"6019939f-3efa-47ee-9570-b9c14cea2ecc\" (UID: \"6019939f-3efa-47ee-9570-b9c14cea2ecc\") " Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.675990 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.676021 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp6lv\" (UniqueName: \"kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.679375 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.687542 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz" (OuterVolumeSpecName: "kube-api-access-7s2sz") pod "6019939f-3efa-47ee-9570-b9c14cea2ecc" (UID: "6019939f-3efa-47ee-9570-b9c14cea2ecc"). InnerVolumeSpecName "kube-api-access-7s2sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.700654 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp6lv\" (UniqueName: \"kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv\") pod \"glance-db-create-znq7k\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " pod="openstack/glance-db-create-znq7k" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.722104 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6019939f-3efa-47ee-9570-b9c14cea2ecc" (UID: "6019939f-3efa-47ee-9570-b9c14cea2ecc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.732902 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6019939f-3efa-47ee-9570-b9c14cea2ecc" (UID: "6019939f-3efa-47ee-9570-b9c14cea2ecc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.738340 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6019939f-3efa-47ee-9570-b9c14cea2ecc" (UID: "6019939f-3efa-47ee-9570-b9c14cea2ecc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.745764 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config" (OuterVolumeSpecName: "config") pod "6019939f-3efa-47ee-9570-b9c14cea2ecc" (UID: "6019939f-3efa-47ee-9570-b9c14cea2ecc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.777442 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.777474 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.777489 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.777502 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s2sz\" (UniqueName: \"kubernetes.io/projected/6019939f-3efa-47ee-9570-b9c14cea2ecc-kube-api-access-7s2sz\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.777515 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6019939f-3efa-47ee-9570-b9c14cea2ecc-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.829362 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" event={"ID":"6019939f-3efa-47ee-9570-b9c14cea2ecc","Type":"ContainerDied","Data":"fcb7c5fe636f87cb55a5d21f5d10236ac83ff9f326e0cd394c459fa7185d6e0c"} Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.829440 4867 scope.go:117] "RemoveContainer" containerID="e7af4d5bd75e5a57bd2b4d0a42ed21234adc4add57ef546e84d2fd53187091fc" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.829499 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-gf5v8" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.833909 4867 generic.go:334] "Generic (PLEG): container finished" podID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerID="942a76f37a7a5e61ee395a3c4846d0cf9ed5d2e8446b251c7f32e5379bd97c85" exitCode=0 Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.833965 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" event={"ID":"b0f75dd4-02a1-473c-a06f-c8477baa2d90","Type":"ContainerDied","Data":"942a76f37a7a5e61ee395a3c4846d0cf9ed5d2e8446b251c7f32e5379bd97c85"} Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.868354 4867 scope.go:117] "RemoveContainer" containerID="ec335c45dc9da9858d39e20d3a99708169d7880067e0cfa31406c514adf32302" Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.898665 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.906918 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-gf5v8"] Dec 12 07:06:37 crc kubenswrapper[4867]: I1212 07:06:37.998486 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-znq7k" Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.147984 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a6d9-account-create-update-s9hgd"] Dec 12 07:06:38 crc kubenswrapper[4867]: W1212 07:06:38.151319 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3ba780e_0b5f_4e6f_9250_2b1eff1cbec3.slice/crio-714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d WatchSource:0}: Error finding container 714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d: Status 404 returned error can't find the container with id 714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.453246 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-znq7k"] Dec 12 07:06:38 crc kubenswrapper[4867]: W1212 07:06:38.455941 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2e4e236_bbc3_42bd_8c73_3386ff4556ba.slice/crio-33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5 WatchSource:0}: Error finding container 33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5: Status 404 returned error can't find the container with id 33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5 Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.846961 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" path="/var/lib/kubelet/pods/6019939f-3efa-47ee-9570-b9c14cea2ecc/volumes" Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.847774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a6d9-account-create-update-s9hgd" event={"ID":"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3","Type":"ContainerStarted","Data":"714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d"} Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.847796 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-znq7k" event={"ID":"d2e4e236-bbc3-42bd-8c73-3386ff4556ba","Type":"ContainerStarted","Data":"33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5"} Dec 12 07:06:38 crc kubenswrapper[4867]: I1212 07:06:38.900815 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:38 crc kubenswrapper[4867]: E1212 07:06:38.901065 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:38 crc kubenswrapper[4867]: E1212 07:06:38.901107 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:38 crc kubenswrapper[4867]: E1212 07:06:38.901175 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:06:42.901153385 +0000 UTC m=+1090.472534654 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.150597 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6nnh4"] Dec 12 07:06:39 crc kubenswrapper[4867]: E1212 07:06:39.151172 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="dnsmasq-dns" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.151188 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="dnsmasq-dns" Dec 12 07:06:39 crc kubenswrapper[4867]: E1212 07:06:39.151206 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="init" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.151212 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="init" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.151402 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6019939f-3efa-47ee-9570-b9c14cea2ecc" containerName="dnsmasq-dns" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.151911 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.156400 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.156805 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.158519 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.193535 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6nnh4"] Dec 12 07:06:39 crc kubenswrapper[4867]: E1212 07:06:39.194108 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-wsw8n ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-wsw8n ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-6nnh4" podUID="7024d8a9-a3b4-47be-b95a-8c49f9896f62" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.218190 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6nnh4"] Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.227990 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bxvfm"] Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.229217 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.247875 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bxvfm"] Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307347 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307395 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sgr5\" (UniqueName: \"kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307430 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307606 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307646 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307678 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307727 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307756 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsw8n\" (UniqueName: \"kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307784 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307863 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.307985 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.308014 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.308029 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.308071 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409298 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsw8n\" (UniqueName: \"kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409387 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409439 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409498 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409517 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409532 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409560 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409582 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409597 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sgr5\" (UniqueName: \"kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409618 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409645 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409662 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.409678 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.410159 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.411405 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.411437 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.412065 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.412170 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.412481 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.415743 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.415851 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.416636 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.416911 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.417574 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.422083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.425879 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsw8n\" (UniqueName: \"kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n\") pod \"swift-ring-rebalance-6nnh4\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.426484 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sgr5\" (UniqueName: \"kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5\") pod \"swift-ring-rebalance-bxvfm\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.556150 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.864534 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a6d9-account-create-update-s9hgd" event={"ID":"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3","Type":"ContainerStarted","Data":"95bf09c837b62e7939b11139065bedf6b217dfd70415ec0d948de99abeef31dc"} Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.868643 4867 generic.go:334] "Generic (PLEG): container finished" podID="d2e4e236-bbc3-42bd-8c73-3386ff4556ba" containerID="ff33b41e33f666b184b75128290757493fa986e1b0ac7b622f277504d200fe71" exitCode=0 Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.869062 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-znq7k" event={"ID":"d2e4e236-bbc3-42bd-8c73-3386ff4556ba","Type":"ContainerDied","Data":"ff33b41e33f666b184b75128290757493fa986e1b0ac7b622f277504d200fe71"} Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.875518 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.876181 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" event={"ID":"b0f75dd4-02a1-473c-a06f-c8477baa2d90","Type":"ContainerStarted","Data":"6cbb9d6eb9e947faca9544cd025ec52e5ab4eabca3e78cfb9f4d07e4d9f9b9b4"} Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.876381 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.888586 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-a6d9-account-create-update-s9hgd" podStartSLOduration=2.88856231 podStartE2EDuration="2.88856231s" podCreationTimestamp="2025-12-12 07:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:39.88034649 +0000 UTC m=+1087.451727759" watchObservedRunningTime="2025-12-12 07:06:39.88856231 +0000 UTC m=+1087.459943569" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.916290 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podStartSLOduration=5.916264442 podStartE2EDuration="5.916264442s" podCreationTimestamp="2025-12-12 07:06:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:39.913824113 +0000 UTC m=+1087.485205382" watchObservedRunningTime="2025-12-12 07:06:39.916264442 +0000 UTC m=+1087.487645721" Dec 12 07:06:39 crc kubenswrapper[4867]: I1212 07:06:39.963548 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.010331 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bxvfm"] Dec 12 07:06:40 crc kubenswrapper[4867]: W1212 07:06:40.012414 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1432b4ca_14f2_4044_a58c_d93ff9930e17.slice/crio-581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6 WatchSource:0}: Error finding container 581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6: Status 404 returned error can't find the container with id 581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6 Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120191 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120321 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120477 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120521 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120542 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsw8n\" (UniqueName: \"kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120608 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.120627 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices\") pod \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\" (UID: \"7024d8a9-a3b4-47be-b95a-8c49f9896f62\") " Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.121254 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.121352 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.122824 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts" (OuterVolumeSpecName: "scripts") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.125654 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.125697 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n" (OuterVolumeSpecName: "kube-api-access-wsw8n") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "kube-api-access-wsw8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.127043 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.132811 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7024d8a9-a3b4-47be-b95a-8c49f9896f62" (UID: "7024d8a9-a3b4-47be-b95a-8c49f9896f62"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222863 4867 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222903 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222912 4867 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222920 4867 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7024d8a9-a3b4-47be-b95a-8c49f9896f62-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222929 4867 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7024d8a9-a3b4-47be-b95a-8c49f9896f62-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222938 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsw8n\" (UniqueName: \"kubernetes.io/projected/7024d8a9-a3b4-47be-b95a-8c49f9896f62-kube-api-access-wsw8n\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.222948 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7024d8a9-a3b4-47be-b95a-8c49f9896f62-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.887357 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" containerID="95bf09c837b62e7939b11139065bedf6b217dfd70415ec0d948de99abeef31dc" exitCode=0 Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.887455 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a6d9-account-create-update-s9hgd" event={"ID":"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3","Type":"ContainerDied","Data":"95bf09c837b62e7939b11139065bedf6b217dfd70415ec0d948de99abeef31dc"} Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.889695 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bxvfm" event={"ID":"1432b4ca-14f2-4044-a58c-d93ff9930e17","Type":"ContainerStarted","Data":"581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6"} Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.889831 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6nnh4" Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.952101 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6nnh4"] Dec 12 07:06:40 crc kubenswrapper[4867]: I1212 07:06:40.961413 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-6nnh4"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.228781 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-znq7k" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.344150 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp6lv\" (UniqueName: \"kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv\") pod \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.344317 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts\") pod \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\" (UID: \"d2e4e236-bbc3-42bd-8c73-3386ff4556ba\") " Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.344851 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2e4e236-bbc3-42bd-8c73-3386ff4556ba" (UID: "d2e4e236-bbc3-42bd-8c73-3386ff4556ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.352395 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv" (OuterVolumeSpecName: "kube-api-access-mp6lv") pod "d2e4e236-bbc3-42bd-8c73-3386ff4556ba" (UID: "d2e4e236-bbc3-42bd-8c73-3386ff4556ba"). InnerVolumeSpecName "kube-api-access-mp6lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.445795 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp6lv\" (UniqueName: \"kubernetes.io/projected/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-kube-api-access-mp6lv\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.445829 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2e4e236-bbc3-42bd-8c73-3386ff4556ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.634816 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-87645"] Dec 12 07:06:41 crc kubenswrapper[4867]: E1212 07:06:41.635688 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e4e236-bbc3-42bd-8c73-3386ff4556ba" containerName="mariadb-database-create" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.635829 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e4e236-bbc3-42bd-8c73-3386ff4556ba" containerName="mariadb-database-create" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.636253 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e4e236-bbc3-42bd-8c73-3386ff4556ba" containerName="mariadb-database-create" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.637130 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.641999 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-87645"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.649813 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-574l2\" (UniqueName: \"kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.650521 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.739985 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8cda-account-create-update-hr2p8"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.741434 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.745078 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.751318 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.751376 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js7cb\" (UniqueName: \"kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.751397 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.751601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-574l2\" (UniqueName: \"kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.752225 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.761857 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cda-account-create-update-hr2p8"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.773674 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-574l2\" (UniqueName: \"kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2\") pod \"keystone-db-create-87645\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " pod="openstack/keystone-db-create-87645" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.852761 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.853057 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js7cb\" (UniqueName: \"kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.853768 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.879926 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js7cb\" (UniqueName: \"kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb\") pod \"keystone-8cda-account-create-update-hr2p8\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.910159 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-znq7k" event={"ID":"d2e4e236-bbc3-42bd-8c73-3386ff4556ba","Type":"ContainerDied","Data":"33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5"} Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.910185 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-znq7k" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.910203 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33edf431e2b7d7bb435f698e67926334e57fcb0f08e04c00c12d90ff86b275c5" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.918939 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-pdtz8"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.920043 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.930684 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pdtz8"] Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.954870 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.954928 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2f6x\" (UniqueName: \"kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:41 crc kubenswrapper[4867]: I1212 07:06:41.955894 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87645" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.057762 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.057958 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2f6x\" (UniqueName: \"kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.059947 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.067988 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.083661 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c9b0-account-create-update-kmdk9"] Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.084341 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2f6x\" (UniqueName: \"kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x\") pod \"placement-db-create-pdtz8\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.085693 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.089873 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.100220 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c9b0-account-create-update-kmdk9"] Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.161388 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.161501 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kb5c\" (UniqueName: \"kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.261871 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.263931 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kb5c\" (UniqueName: \"kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.264054 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.265027 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.283022 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.290028 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kb5c\" (UniqueName: \"kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c\") pod \"placement-c9b0-account-create-update-kmdk9\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.416823 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.466498 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5b9j\" (UniqueName: \"kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j\") pod \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.466641 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts\") pod \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\" (UID: \"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3\") " Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.468427 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" (UID: "a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.472660 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j" (OuterVolumeSpecName: "kube-api-access-f5b9j") pod "a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" (UID: "a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3"). InnerVolumeSpecName "kube-api-access-f5b9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.569777 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5b9j\" (UniqueName: \"kubernetes.io/projected/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-kube-api-access-f5b9j\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.569804 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.582919 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-87645"] Dec 12 07:06:42 crc kubenswrapper[4867]: W1212 07:06:42.588635 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6b7070b_03c8_4d2b_b72a_6474d1e30cc2.slice/crio-c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b WatchSource:0}: Error finding container c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b: Status 404 returned error can't find the container with id c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.677687 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cda-account-create-update-hr2p8"] Dec 12 07:06:42 crc kubenswrapper[4867]: W1212 07:06:42.709472 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83f68187_8afa_443c_bce0_1d5aff2a234e.slice/crio-fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061 WatchSource:0}: Error finding container fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061: Status 404 returned error can't find the container with id fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061 Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.782268 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-pdtz8"] Dec 12 07:06:42 crc kubenswrapper[4867]: W1212 07:06:42.806961 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e8dc60f_09fa_4c84_b22d_1014fc6d39b4.slice/crio-5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3 WatchSource:0}: Error finding container 5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3: Status 404 returned error can't find the container with id 5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3 Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.851613 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7024d8a9-a3b4-47be-b95a-8c49f9896f62" path="/var/lib/kubelet/pods/7024d8a9-a3b4-47be-b95a-8c49f9896f62/volumes" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.898085 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c9b0-account-create-update-kmdk9"] Dec 12 07:06:42 crc kubenswrapper[4867]: W1212 07:06:42.916316 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcba96708_8c06_4d68_bd49_1cf930cc5cfb.slice/crio-9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6 WatchSource:0}: Error finding container 9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6: Status 404 returned error can't find the container with id 9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6 Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.932830 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a6d9-account-create-update-s9hgd" event={"ID":"a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3","Type":"ContainerDied","Data":"714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.932867 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a6d9-account-create-update-s9hgd" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.932887 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="714c16453d85358b5c82c19eeeff9a1cc2bf9a08d8cdb8d46a80e65b59fdd59d" Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.935945 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pdtz8" event={"ID":"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4","Type":"ContainerStarted","Data":"5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.938439 4867 generic.go:334] "Generic (PLEG): container finished" podID="00532992-54f5-4a24-b0e4-eb5512e08753" containerID="0f947d9d4fe9c6b4d937ad0d0687b9d1db4c6172ed3a4aac66489775077c3e68" exitCode=0 Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.938521 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerDied","Data":"0f947d9d4fe9c6b4d937ad0d0687b9d1db4c6172ed3a4aac66489775077c3e68"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.947065 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87645" event={"ID":"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2","Type":"ContainerStarted","Data":"04e9eda5d1420cdeac04a51995e5de475c02ac1d5311c054f1124ae87b3fb8f0"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.947118 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87645" event={"ID":"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2","Type":"ContainerStarted","Data":"c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.951332 4867 generic.go:334] "Generic (PLEG): container finished" podID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerID="de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020" exitCode=0 Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.951414 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerDied","Data":"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.953802 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cda-account-create-update-hr2p8" event={"ID":"83f68187-8afa-443c-bce0-1d5aff2a234e","Type":"ContainerStarted","Data":"30179a07ae8844a4ffe7aef6f4c96372c3ae3317a61f76efde0b33fbe5dc214c"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.953830 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cda-account-create-update-hr2p8" event={"ID":"83f68187-8afa-443c-bce0-1d5aff2a234e","Type":"ContainerStarted","Data":"fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061"} Dec 12 07:06:42 crc kubenswrapper[4867]: I1212 07:06:42.977405 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:42 crc kubenswrapper[4867]: E1212 07:06:42.978441 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:42 crc kubenswrapper[4867]: E1212 07:06:42.978462 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:42 crc kubenswrapper[4867]: E1212 07:06:42.978588 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:06:50.978488482 +0000 UTC m=+1098.549869751 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.010353 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8cda-account-create-update-hr2p8" podStartSLOduration=2.010333385 podStartE2EDuration="2.010333385s" podCreationTimestamp="2025-12-12 07:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:42.97552087 +0000 UTC m=+1090.546902149" watchObservedRunningTime="2025-12-12 07:06:43.010333385 +0000 UTC m=+1090.581714674" Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.016965 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-87645" podStartSLOduration=2.016946016 podStartE2EDuration="2.016946016s" podCreationTimestamp="2025-12-12 07:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:06:42.994147343 +0000 UTC m=+1090.565528622" watchObservedRunningTime="2025-12-12 07:06:43.016946016 +0000 UTC m=+1090.588327285" Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.967845 4867 generic.go:334] "Generic (PLEG): container finished" podID="cba96708-8c06-4d68-bd49-1cf930cc5cfb" containerID="8c81c50804367f392fba39b977b26bfecdb90e8e5eb09c175d3b03b46dbc1469" exitCode=0 Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.967895 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c9b0-account-create-update-kmdk9" event={"ID":"cba96708-8c06-4d68-bd49-1cf930cc5cfb","Type":"ContainerDied","Data":"8c81c50804367f392fba39b977b26bfecdb90e8e5eb09c175d3b03b46dbc1469"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.968254 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c9b0-account-create-update-kmdk9" event={"ID":"cba96708-8c06-4d68-bd49-1cf930cc5cfb","Type":"ContainerStarted","Data":"9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.972584 4867 generic.go:334] "Generic (PLEG): container finished" podID="2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" containerID="61dac5c9e6307be96fa44d7a6e3d7fc40e109ecc80d7a7042569d594f2df0414" exitCode=0 Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.972941 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pdtz8" event={"ID":"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4","Type":"ContainerDied","Data":"61dac5c9e6307be96fa44d7a6e3d7fc40e109ecc80d7a7042569d594f2df0414"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.977468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerStarted","Data":"02be4e6b1d82467e195a35fda03f0c2a9e4fbba01b3f18526ddc7beb6bb54785"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.977709 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.979832 4867 generic.go:334] "Generic (PLEG): container finished" podID="b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" containerID="04e9eda5d1420cdeac04a51995e5de475c02ac1d5311c054f1124ae87b3fb8f0" exitCode=0 Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.979870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87645" event={"ID":"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2","Type":"ContainerDied","Data":"04e9eda5d1420cdeac04a51995e5de475c02ac1d5311c054f1124ae87b3fb8f0"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.982256 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerStarted","Data":"234415f3af38fbe83a304c1655d565aaee048e266eab8c74960cba70dcab3267"} Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.982451 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.983978 4867 generic.go:334] "Generic (PLEG): container finished" podID="83f68187-8afa-443c-bce0-1d5aff2a234e" containerID="30179a07ae8844a4ffe7aef6f4c96372c3ae3317a61f76efde0b33fbe5dc214c" exitCode=0 Dec 12 07:06:43 crc kubenswrapper[4867]: I1212 07:06:43.984014 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cda-account-create-update-hr2p8" event={"ID":"83f68187-8afa-443c-bce0-1d5aff2a234e","Type":"ContainerDied","Data":"30179a07ae8844a4ffe7aef6f4c96372c3ae3317a61f76efde0b33fbe5dc214c"} Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.023780 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.961834521 podStartE2EDuration="58.023764661s" podCreationTimestamp="2025-12-12 07:05:46 +0000 UTC" firstStartedPulling="2025-12-12 07:05:57.224261056 +0000 UTC m=+1044.795642325" lastFinishedPulling="2025-12-12 07:06:09.286191196 +0000 UTC m=+1056.857572465" observedRunningTime="2025-12-12 07:06:44.020680866 +0000 UTC m=+1091.592062135" watchObservedRunningTime="2025-12-12 07:06:44.023764661 +0000 UTC m=+1091.595145930" Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.075829 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.953479569 podStartE2EDuration="57.075812035s" podCreationTimestamp="2025-12-12 07:05:47 +0000 UTC" firstStartedPulling="2025-12-12 07:05:57.226741377 +0000 UTC m=+1044.798122646" lastFinishedPulling="2025-12-12 07:06:09.349073843 +0000 UTC m=+1056.920455112" observedRunningTime="2025-12-12 07:06:44.070969627 +0000 UTC m=+1091.642350906" watchObservedRunningTime="2025-12-12 07:06:44.075812035 +0000 UTC m=+1091.647193304" Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.355076 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.421902 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.422189 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="dnsmasq-dns" containerID="cri-o://20a113b51f1390fced95e76529377d456fd6c7eb03779413cec0fa70965c1ac2" gracePeriod=10 Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.995325 4867 generic.go:334] "Generic (PLEG): container finished" podID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerID="20a113b51f1390fced95e76529377d456fd6c7eb03779413cec0fa70965c1ac2" exitCode=0 Dec 12 07:06:44 crc kubenswrapper[4867]: I1212 07:06:44.996415 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" event={"ID":"87d29eb9-e24f-417c-b71f-51bbb63d3a58","Type":"ContainerDied","Data":"20a113b51f1390fced95e76529377d456fd6c7eb03779413cec0fa70965c1ac2"} Dec 12 07:06:45 crc kubenswrapper[4867]: I1212 07:06:45.456149 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.165682 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87645" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.177045 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.189074 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.198726 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260513 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js7cb\" (UniqueName: \"kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb\") pod \"83f68187-8afa-443c-bce0-1d5aff2a234e\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260614 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts\") pod \"83f68187-8afa-443c-bce0-1d5aff2a234e\" (UID: \"83f68187-8afa-443c-bce0-1d5aff2a234e\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260768 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kb5c\" (UniqueName: \"kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c\") pod \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260854 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2f6x\" (UniqueName: \"kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x\") pod \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260890 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts\") pod \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\" (UID: \"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260928 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts\") pod \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\" (UID: \"cba96708-8c06-4d68-bd49-1cf930cc5cfb\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.260984 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts\") pod \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261027 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-574l2\" (UniqueName: \"kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2\") pod \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\" (UID: \"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2\") " Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261295 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83f68187-8afa-443c-bce0-1d5aff2a234e" (UID: "83f68187-8afa-443c-bce0-1d5aff2a234e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261434 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" (UID: "2e8dc60f-09fa-4c84-b22d-1014fc6d39b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261796 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" (UID: "b6b7070b-03c8-4d2b-b72a-6474d1e30cc2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261828 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cba96708-8c06-4d68-bd49-1cf930cc5cfb" (UID: "cba96708-8c06-4d68-bd49-1cf930cc5cfb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261878 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.261898 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83f68187-8afa-443c-bce0-1d5aff2a234e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.269684 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x" (OuterVolumeSpecName: "kube-api-access-s2f6x") pod "2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" (UID: "2e8dc60f-09fa-4c84-b22d-1014fc6d39b4"). InnerVolumeSpecName "kube-api-access-s2f6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.269745 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb" (OuterVolumeSpecName: "kube-api-access-js7cb") pod "83f68187-8afa-443c-bce0-1d5aff2a234e" (UID: "83f68187-8afa-443c-bce0-1d5aff2a234e"). InnerVolumeSpecName "kube-api-access-js7cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.275443 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c" (OuterVolumeSpecName: "kube-api-access-5kb5c") pod "cba96708-8c06-4d68-bd49-1cf930cc5cfb" (UID: "cba96708-8c06-4d68-bd49-1cf930cc5cfb"). InnerVolumeSpecName "kube-api-access-5kb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.295574 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2" (OuterVolumeSpecName: "kube-api-access-574l2") pod "b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" (UID: "b6b7070b-03c8-4d2b-b72a-6474d1e30cc2"). InnerVolumeSpecName "kube-api-access-574l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364754 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2f6x\" (UniqueName: \"kubernetes.io/projected/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4-kube-api-access-s2f6x\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364818 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cba96708-8c06-4d68-bd49-1cf930cc5cfb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364828 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364838 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-574l2\" (UniqueName: \"kubernetes.io/projected/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2-kube-api-access-574l2\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364847 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js7cb\" (UniqueName: \"kubernetes.io/projected/83f68187-8afa-443c-bce0-1d5aff2a234e-kube-api-access-js7cb\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.364857 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kb5c\" (UniqueName: \"kubernetes.io/projected/cba96708-8c06-4d68-bd49-1cf930cc5cfb-kube-api-access-5kb5c\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.484411 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-7lxjs"] Dec 12 07:06:47 crc kubenswrapper[4867]: E1212 07:06:47.485226 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f68187-8afa-443c-bce0-1d5aff2a234e" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485266 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f68187-8afa-443c-bce0-1d5aff2a234e" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: E1212 07:06:47.485280 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba96708-8c06-4d68-bd49-1cf930cc5cfb" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485291 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba96708-8c06-4d68-bd49-1cf930cc5cfb" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: E1212 07:06:47.485306 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485317 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: E1212 07:06:47.485341 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485352 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: E1212 07:06:47.485385 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485395 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485593 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f68187-8afa-443c-bce0-1d5aff2a234e" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485610 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485628 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" containerName="mariadb-database-create" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485639 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba96708-8c06-4d68-bd49-1cf930cc5cfb" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.485656 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" containerName="mariadb-account-create-update" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.486346 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.488774 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.488820 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lggbv" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.504761 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7lxjs"] Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.567004 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.567073 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbndn\" (UniqueName: \"kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.567407 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.567609 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.668818 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.668942 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.668987 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.669016 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbndn\" (UniqueName: \"kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.673363 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.673833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.675367 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.685284 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbndn\" (UniqueName: \"kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn\") pod \"glance-db-sync-7lxjs\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.805107 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7lxjs" Dec 12 07:06:47 crc kubenswrapper[4867]: I1212 07:06:47.876119 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.98:5353: connect: connection refused" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.023773 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-87645" event={"ID":"b6b7070b-03c8-4d2b-b72a-6474d1e30cc2","Type":"ContainerDied","Data":"c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b"} Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.023825 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38ebf1cf54cd2f1add8112636e8103b4cc99283f069030a80baa1c6c069854b" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.023795 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-87645" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.025769 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cda-account-create-update-hr2p8" event={"ID":"83f68187-8afa-443c-bce0-1d5aff2a234e","Type":"ContainerDied","Data":"fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061"} Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.025801 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe7473c741cadc6c064a5285610a558d55cab4a6ebd3ceea0f9f52201c54b061" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.025799 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cda-account-create-update-hr2p8" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.027773 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c9b0-account-create-update-kmdk9" event={"ID":"cba96708-8c06-4d68-bd49-1cf930cc5cfb","Type":"ContainerDied","Data":"9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6"} Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.027834 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9378e562ee5dc85df942be13569b784006d4195c1602502d21b8e1b4155bb1d6" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.027819 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c9b0-account-create-update-kmdk9" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.050647 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-pdtz8" event={"ID":"2e8dc60f-09fa-4c84-b22d-1014fc6d39b4","Type":"ContainerDied","Data":"5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3"} Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.050686 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d3191e3915000b6309d526798e51912fed71f07c9454657d987e7a67051c4e3" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.050750 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-pdtz8" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.300165 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.379909 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc\") pod \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.380049 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtx8h\" (UniqueName: \"kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h\") pod \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.380167 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config\") pod \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\" (UID: \"87d29eb9-e24f-417c-b71f-51bbb63d3a58\") " Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.399490 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h" (OuterVolumeSpecName: "kube-api-access-mtx8h") pod "87d29eb9-e24f-417c-b71f-51bbb63d3a58" (UID: "87d29eb9-e24f-417c-b71f-51bbb63d3a58"). InnerVolumeSpecName "kube-api-access-mtx8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.432181 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7lxjs"] Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.439293 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config" (OuterVolumeSpecName: "config") pod "87d29eb9-e24f-417c-b71f-51bbb63d3a58" (UID: "87d29eb9-e24f-417c-b71f-51bbb63d3a58"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.452218 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87d29eb9-e24f-417c-b71f-51bbb63d3a58" (UID: "87d29eb9-e24f-417c-b71f-51bbb63d3a58"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.482055 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.482088 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87d29eb9-e24f-417c-b71f-51bbb63d3a58-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:48 crc kubenswrapper[4867]: I1212 07:06:48.482098 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtx8h\" (UniqueName: \"kubernetes.io/projected/87d29eb9-e24f-417c-b71f-51bbb63d3a58-kube-api-access-mtx8h\") on node \"crc\" DevicePath \"\"" Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.075967 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7lxjs" event={"ID":"a3a5237b-ffff-43b1-a921-68affabec6cd","Type":"ContainerStarted","Data":"62708c7aca8968aaf36f542b839b98f234c61924be7fa50ef39ff756fb422dd3"} Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.080285 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" event={"ID":"87d29eb9-e24f-417c-b71f-51bbb63d3a58","Type":"ContainerDied","Data":"2917adab8cd8fc51b4fe0bbebd42cee593c1b4c4eca865dc2f6bf169f51c60dd"} Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.080409 4867 scope.go:117] "RemoveContainer" containerID="20a113b51f1390fced95e76529377d456fd6c7eb03779413cec0fa70965c1ac2" Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.080559 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-rslfd" Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.083726 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bxvfm" event={"ID":"1432b4ca-14f2-4044-a58c-d93ff9930e17","Type":"ContainerStarted","Data":"651d2a085bb2c2fbadc12806221bb243108cd7ac6a3dcb3f1154eed3ff0a88ed"} Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.112782 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bxvfm" podStartSLOduration=1.596078437 podStartE2EDuration="10.11276068s" podCreationTimestamp="2025-12-12 07:06:39 +0000 UTC" firstStartedPulling="2025-12-12 07:06:40.019779405 +0000 UTC m=+1087.591160674" lastFinishedPulling="2025-12-12 07:06:48.536461648 +0000 UTC m=+1096.107842917" observedRunningTime="2025-12-12 07:06:49.108726562 +0000 UTC m=+1096.680107841" watchObservedRunningTime="2025-12-12 07:06:49.11276068 +0000 UTC m=+1096.684141939" Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.128071 4867 scope.go:117] "RemoveContainer" containerID="f17004fdeeb584465b109ce2494ff8b3ebe6be7759298ba2e3617b73e515a789" Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.135855 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:06:49 crc kubenswrapper[4867]: I1212 07:06:49.142922 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-rslfd"] Dec 12 07:06:50 crc kubenswrapper[4867]: I1212 07:06:50.848807 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" path="/var/lib/kubelet/pods/87d29eb9-e24f-417c-b71f-51bbb63d3a58/volumes" Dec 12 07:06:51 crc kubenswrapper[4867]: I1212 07:06:51.025351 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:06:51 crc kubenswrapper[4867]: E1212 07:06:51.025689 4867 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 12 07:06:51 crc kubenswrapper[4867]: E1212 07:06:51.025708 4867 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 12 07:06:51 crc kubenswrapper[4867]: E1212 07:06:51.025768 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift podName:786b1010-d0a8-4c94-a42c-99226ab568be nodeName:}" failed. No retries permitted until 2025-12-12 07:07:07.025752857 +0000 UTC m=+1114.597134126 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift") pod "swift-storage-0" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be") : configmap "swift-ring-files" not found Dec 12 07:06:52 crc kubenswrapper[4867]: I1212 07:06:52.320304 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" probeResult="failure" output=< Dec 12 07:06:52 crc kubenswrapper[4867]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 12 07:06:52 crc kubenswrapper[4867]: > Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.320174 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" probeResult="failure" output=< Dec 12 07:06:57 crc kubenswrapper[4867]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 12 07:06:57 crc kubenswrapper[4867]: > Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.345819 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.360491 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.626430 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8njtj-config-b9p6s"] Dec 12 07:06:57 crc kubenswrapper[4867]: E1212 07:06:57.627059 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="dnsmasq-dns" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.627166 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="dnsmasq-dns" Dec 12 07:06:57 crc kubenswrapper[4867]: E1212 07:06:57.627246 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="init" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.627314 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="init" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.627542 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="87d29eb9-e24f-417c-b71f-51bbb63d3a58" containerName="dnsmasq-dns" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.628201 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.635535 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642200 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvgtw\" (UniqueName: \"kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642501 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642659 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642752 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642892 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.642992 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.649207 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj-config-b9p6s"] Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744447 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744510 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744546 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744575 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744618 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvgtw\" (UniqueName: \"kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744659 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.744997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.745032 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.745877 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.746092 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.747051 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.767935 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvgtw\" (UniqueName: \"kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw\") pod \"ovn-controller-8njtj-config-b9p6s\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:57 crc kubenswrapper[4867]: I1212 07:06:57.970511 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:06:59 crc kubenswrapper[4867]: I1212 07:06:59.171669 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:06:59 crc kubenswrapper[4867]: I1212 07:06:59.489536 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.206861 4867 generic.go:334] "Generic (PLEG): container finished" podID="1432b4ca-14f2-4044-a58c-d93ff9930e17" containerID="651d2a085bb2c2fbadc12806221bb243108cd7ac6a3dcb3f1154eed3ff0a88ed" exitCode=0 Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.206947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bxvfm" event={"ID":"1432b4ca-14f2-4044-a58c-d93ff9930e17","Type":"ContainerDied","Data":"651d2a085bb2c2fbadc12806221bb243108cd7ac6a3dcb3f1154eed3ff0a88ed"} Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.348292 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qtct9"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.350111 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.359799 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qtct9"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.411936 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-x8h7k"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.413286 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.418390 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.418461 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78ggr\" (UniqueName: \"kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.431030 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x8h7k"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.518311 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b8be-account-create-update-29zkz"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.520122 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.520521 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.520574 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78ggr\" (UniqueName: \"kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.520620 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh5mp\" (UniqueName: \"kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.520649 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.521695 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.527729 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b8be-account-create-update-29zkz"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.532346 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.564274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78ggr\" (UniqueName: \"kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr\") pod \"cinder-db-create-qtct9\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.621823 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh5mp\" (UniqueName: \"kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.621904 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.621963 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.622050 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqk7r\" (UniqueName: \"kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.623145 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.632051 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-74a1-account-create-update-mdf7g"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.633546 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.637407 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.648038 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-74a1-account-create-update-mdf7g"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.671965 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.678982 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh5mp\" (UniqueName: \"kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp\") pod \"barbican-db-create-x8h7k\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.692308 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-n2xp2"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.693781 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.699940 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9jg9" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.700211 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.700495 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.700805 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.717127 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-n2xp2"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.726149 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqk7r\" (UniqueName: \"kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.726212 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkpv7\" (UniqueName: \"kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.726272 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.728794 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.728886 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.728981 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmxhr\" (UniqueName: \"kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.729071 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.732595 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.746612 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.756894 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-dz67c"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.758839 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.764363 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqk7r\" (UniqueName: \"kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r\") pod \"cinder-b8be-account-create-update-29zkz\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.791896 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dz67c"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.830692 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkpv7\" (UniqueName: \"kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.831450 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.831629 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.831873 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqkf2\" (UniqueName: \"kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.832037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.832195 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmxhr\" (UniqueName: \"kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.832479 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.833534 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.836484 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.838324 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.845292 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.847866 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkpv7\" (UniqueName: \"kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7\") pod \"barbican-74a1-account-create-update-mdf7g\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.853168 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmxhr\" (UniqueName: \"kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr\") pod \"keystone-db-sync-n2xp2\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.918119 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-29f2-account-create-update-qfrmt"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.919593 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.924164 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.934178 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c7b9\" (UniqueName: \"kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.934292 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.934376 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.935248 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.935638 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqkf2\" (UniqueName: \"kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.948908 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-29f2-account-create-update-qfrmt"] Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.963754 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqkf2\" (UniqueName: \"kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2\") pod \"neutron-db-create-dz67c\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:01 crc kubenswrapper[4867]: I1212 07:07:01.968658 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.037936 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c7b9\" (UniqueName: \"kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.038349 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.039482 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.053107 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.061990 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c7b9\" (UniqueName: \"kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9\") pod \"neutron-29f2-account-create-update-qfrmt\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.121689 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.254063 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.341581 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" probeResult="failure" output=< Dec 12 07:07:02 crc kubenswrapper[4867]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 12 07:07:02 crc kubenswrapper[4867]: > Dec 12 07:07:02 crc kubenswrapper[4867]: I1212 07:07:02.897701 4867 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf3346b14-1b14-4a53-953b-5c6a2d6ff724"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf3346b14-1b14-4a53-953b-5c6a2d6ff724] : Timed out while waiting for systemd to remove kubepods-besteffort-podf3346b14_1b14_4a53_953b_5c6a2d6ff724.slice" Dec 12 07:07:02 crc kubenswrapper[4867]: E1212 07:07:02.898099 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podf3346b14-1b14-4a53-953b-5c6a2d6ff724] : unable to destroy cgroup paths for cgroup [kubepods besteffort podf3346b14-1b14-4a53-953b-5c6a2d6ff724] : Timed out while waiting for systemd to remove kubepods-besteffort-podf3346b14_1b14_4a53_953b_5c6a2d6ff724.slice" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" podUID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" Dec 12 07:07:03 crc kubenswrapper[4867]: I1212 07:07:03.229942 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79764b65-79vhd" Dec 12 07:07:03 crc kubenswrapper[4867]: I1212 07:07:03.337499 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:07:03 crc kubenswrapper[4867]: I1212 07:07:03.355184 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b79764b65-79vhd"] Dec 12 07:07:04 crc kubenswrapper[4867]: E1212 07:07:04.739838 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f" Dec 12 07:07:04 crc kubenswrapper[4867]: E1212 07:07:04.740175 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bbndn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-7lxjs_openstack(a3a5237b-ffff-43b1-a921-68affabec6cd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:07:04 crc kubenswrapper[4867]: E1212 07:07:04.741751 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-7lxjs" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.803032 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.860484 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3346b14-1b14-4a53-953b-5c6a2d6ff724" path="/var/lib/kubelet/pods/f3346b14-1b14-4a53-953b-5c6a2d6ff724/volumes" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894491 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sgr5\" (UniqueName: \"kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894580 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894670 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894738 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894776 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894816 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.894854 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices\") pod \"1432b4ca-14f2-4044-a58c-d93ff9930e17\" (UID: \"1432b4ca-14f2-4044-a58c-d93ff9930e17\") " Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.897100 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.897633 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.902784 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5" (OuterVolumeSpecName: "kube-api-access-8sgr5") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "kube-api-access-8sgr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.905501 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.931757 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts" (OuterVolumeSpecName: "scripts") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.963558 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.971966 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1432b4ca-14f2-4044-a58c-d93ff9930e17" (UID: "1432b4ca-14f2-4044-a58c-d93ff9930e17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998809 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998848 4867 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1432b4ca-14f2-4044-a58c-d93ff9930e17-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998862 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998878 4867 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1432b4ca-14f2-4044-a58c-d93ff9930e17-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998894 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sgr5\" (UniqueName: \"kubernetes.io/projected/1432b4ca-14f2-4044-a58c-d93ff9930e17-kube-api-access-8sgr5\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998908 4867 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:04 crc kubenswrapper[4867]: I1212 07:07:04.998922 4867 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1432b4ca-14f2-4044-a58c-d93ff9930e17-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.233611 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x8h7k"] Dec 12 07:07:05 crc kubenswrapper[4867]: W1212 07:07:05.239374 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8faa97c_b754_433d_ae7c_d875a0cd286e.slice/crio-cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37 WatchSource:0}: Error finding container cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37: Status 404 returned error can't find the container with id cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37 Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.249054 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bxvfm" Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.255322 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bxvfm" event={"ID":"1432b4ca-14f2-4044-a58c-d93ff9930e17","Type":"ContainerDied","Data":"581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6"} Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.255411 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="581bd42f3adf6a62d31102858b3deef4e536a3a15e10cbbdf7674158dfcac3f6" Dec 12 07:07:05 crc kubenswrapper[4867]: E1212 07:07:05.258218 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:e4aa4ebbb1e581a12040e9ad2ae2709ac31b5d965bb64fc4252d1028b05c565f\\\"\"" pod="openstack/glance-db-sync-7lxjs" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.403368 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b8be-account-create-update-29zkz"] Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.475112 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-29f2-account-create-update-qfrmt"] Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.488490 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-n2xp2"] Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.515463 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj-config-b9p6s"] Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.526537 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-74a1-account-create-update-mdf7g"] Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.697394 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qtct9"] Dec 12 07:07:05 crc kubenswrapper[4867]: W1212 07:07:05.702681 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6da4f5f_8f05_47d2_a3d5_79cfd7666d51.slice/crio-437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3 WatchSource:0}: Error finding container 437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3: Status 404 returned error can't find the container with id 437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3 Dec 12 07:07:05 crc kubenswrapper[4867]: W1212 07:07:05.728677 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9fabde1_97a0_4d14_8613_6bd031ab968d.slice/crio-85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58 WatchSource:0}: Error finding container 85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58: Status 404 returned error can't find the container with id 85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58 Dec 12 07:07:05 crc kubenswrapper[4867]: I1212 07:07:05.735040 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dz67c"] Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.263067 4867 generic.go:334] "Generic (PLEG): container finished" podID="5986e88e-5488-4faa-829f-eebfbf81f209" containerID="c5dab6101821935c5fac66af5d8442253153302d4354001cf035fc6582565ef1" exitCode=0 Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.263127 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b8be-account-create-update-29zkz" event={"ID":"5986e88e-5488-4faa-829f-eebfbf81f209","Type":"ContainerDied","Data":"c5dab6101821935c5fac66af5d8442253153302d4354001cf035fc6582565ef1"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.263307 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b8be-account-create-update-29zkz" event={"ID":"5986e88e-5488-4faa-829f-eebfbf81f209","Type":"ContainerStarted","Data":"25ab6bfd724a0c73d90b8c04e67e8060e8751b30650dd0ca96adcbb3f04d0bf8"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.268042 4867 generic.go:334] "Generic (PLEG): container finished" podID="c8faa97c-b754-433d-ae7c-d875a0cd286e" containerID="dea0997833a5437fd02115c35913afaf141ffd915e7a817d66c7a6536e56241f" exitCode=0 Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.268120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x8h7k" event={"ID":"c8faa97c-b754-433d-ae7c-d875a0cd286e","Type":"ContainerDied","Data":"dea0997833a5437fd02115c35913afaf141ffd915e7a817d66c7a6536e56241f"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.268147 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x8h7k" event={"ID":"c8faa97c-b754-433d-ae7c-d875a0cd286e","Type":"ContainerStarted","Data":"cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.270482 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n2xp2" event={"ID":"d930b18c-8189-43e2-b59b-9b75e0ac338b","Type":"ContainerStarted","Data":"644ce48e06096cb81c11ec4f19ce0e17dbdce9809b491e58d8e8df0f0967120d"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.272837 4867 generic.go:334] "Generic (PLEG): container finished" podID="8dcfd4ee-cde3-49e8-ad50-af6fd23af517" containerID="7a47e43be70ffc5bba10ae4e7296ef0e5296adf16d60fc3a9122ae9a7df19a71" exitCode=0 Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.272907 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29f2-account-create-update-qfrmt" event={"ID":"8dcfd4ee-cde3-49e8-ad50-af6fd23af517","Type":"ContainerDied","Data":"7a47e43be70ffc5bba10ae4e7296ef0e5296adf16d60fc3a9122ae9a7df19a71"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.272931 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29f2-account-create-update-qfrmt" event={"ID":"8dcfd4ee-cde3-49e8-ad50-af6fd23af517","Type":"ContainerStarted","Data":"69cbe5b41f9ed22afdfe946e8e8e3017a70ebb2a0c87ee5d126d6ce2d1fe4c61"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.274740 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-b9p6s" event={"ID":"a86271f9-3990-46b9-971d-0151ddda839f","Type":"ContainerStarted","Data":"76fd69aa3c2d1878f209b6a29c4102f2c3ad87c59c62eb8795673bc33e0dc2e5"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.281633 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qtct9" event={"ID":"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51","Type":"ContainerStarted","Data":"62cb32eb7c19228623a0cc83eac79f7834312e294015b665c1c83d3000645c7f"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.281714 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qtct9" event={"ID":"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51","Type":"ContainerStarted","Data":"437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.283976 4867 generic.go:334] "Generic (PLEG): container finished" podID="7ec53377-bb33-44d4-af94-7b24177ff046" containerID="01bac50358ef03d8b64fd11175d1911fa871e096ed6fc433023c26d56989cc80" exitCode=0 Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.284051 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-74a1-account-create-update-mdf7g" event={"ID":"7ec53377-bb33-44d4-af94-7b24177ff046","Type":"ContainerDied","Data":"01bac50358ef03d8b64fd11175d1911fa871e096ed6fc433023c26d56989cc80"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.284081 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-74a1-account-create-update-mdf7g" event={"ID":"7ec53377-bb33-44d4-af94-7b24177ff046","Type":"ContainerStarted","Data":"eb06b723ee7d265611fd63c4a317e53a4c6a41a51c16466b04d5110f87218818"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.285736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dz67c" event={"ID":"b9fabde1-97a0-4d14-8613-6bd031ab968d","Type":"ContainerStarted","Data":"85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58"} Dec 12 07:07:06 crc kubenswrapper[4867]: I1212 07:07:06.346465 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qtct9" podStartSLOduration=5.346442895 podStartE2EDuration="5.346442895s" podCreationTimestamp="2025-12-12 07:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:06.337994497 +0000 UTC m=+1113.909375806" watchObservedRunningTime="2025-12-12 07:07:06.346442895 +0000 UTC m=+1113.917824164" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.060728 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.081777 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"swift-storage-0\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " pod="openstack/swift-storage-0" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.297408 4867 generic.go:334] "Generic (PLEG): container finished" podID="e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" containerID="62cb32eb7c19228623a0cc83eac79f7834312e294015b665c1c83d3000645c7f" exitCode=0 Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.297494 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qtct9" event={"ID":"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51","Type":"ContainerDied","Data":"62cb32eb7c19228623a0cc83eac79f7834312e294015b665c1c83d3000645c7f"} Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.301063 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dz67c" event={"ID":"b9fabde1-97a0-4d14-8613-6bd031ab968d","Type":"ContainerStarted","Data":"77c9a26c17603a8e3b6947d790bbaaaac80e35891cbed056027c6b7726af2e06"} Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.303969 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-b9p6s" event={"ID":"a86271f9-3990-46b9-971d-0151ddda839f","Type":"ContainerStarted","Data":"e951181ca2240c8345911e4e0b3c224a258ab5437f10f82e6d47edbc589d9fb2"} Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.318883 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.335478 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8njtj" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.338549 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8njtj-config-b9p6s" podStartSLOduration=10.338523583 podStartE2EDuration="10.338523583s" podCreationTimestamp="2025-12-12 07:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:07.331357407 +0000 UTC m=+1114.902738706" watchObservedRunningTime="2025-12-12 07:07:07.338523583 +0000 UTC m=+1114.909904852" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.362509 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-dz67c" podStartSLOduration=6.362489331 podStartE2EDuration="6.362489331s" podCreationTimestamp="2025-12-12 07:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:07.349730388 +0000 UTC m=+1114.921111657" watchObservedRunningTime="2025-12-12 07:07:07.362489331 +0000 UTC m=+1114.933870600" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.800754 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.883265 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c7b9\" (UniqueName: \"kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9\") pod \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.883330 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts\") pod \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\" (UID: \"8dcfd4ee-cde3-49e8-ad50-af6fd23af517\") " Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.884507 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8dcfd4ee-cde3-49e8-ad50-af6fd23af517" (UID: "8dcfd4ee-cde3-49e8-ad50-af6fd23af517"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.892342 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9" (OuterVolumeSpecName: "kube-api-access-7c7b9") pod "8dcfd4ee-cde3-49e8-ad50-af6fd23af517" (UID: "8dcfd4ee-cde3-49e8-ad50-af6fd23af517"). InnerVolumeSpecName "kube-api-access-7c7b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.976067 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.988038 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c7b9\" (UniqueName: \"kubernetes.io/projected/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-kube-api-access-7c7b9\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:07 crc kubenswrapper[4867]: I1212 07:07:07.988079 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8dcfd4ee-cde3-49e8-ad50-af6fd23af517-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.018181 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.019927 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089123 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts\") pod \"c8faa97c-b754-433d-ae7c-d875a0cd286e\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089197 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqk7r\" (UniqueName: \"kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r\") pod \"5986e88e-5488-4faa-829f-eebfbf81f209\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089473 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts\") pod \"7ec53377-bb33-44d4-af94-7b24177ff046\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089522 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh5mp\" (UniqueName: \"kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp\") pod \"c8faa97c-b754-433d-ae7c-d875a0cd286e\" (UID: \"c8faa97c-b754-433d-ae7c-d875a0cd286e\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089559 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts\") pod \"5986e88e-5488-4faa-829f-eebfbf81f209\" (UID: \"5986e88e-5488-4faa-829f-eebfbf81f209\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089603 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkpv7\" (UniqueName: \"kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7\") pod \"7ec53377-bb33-44d4-af94-7b24177ff046\" (UID: \"7ec53377-bb33-44d4-af94-7b24177ff046\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.089831 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8faa97c-b754-433d-ae7c-d875a0cd286e" (UID: "c8faa97c-b754-433d-ae7c-d875a0cd286e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.090069 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8faa97c-b754-433d-ae7c-d875a0cd286e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.090709 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ec53377-bb33-44d4-af94-7b24177ff046" (UID: "7ec53377-bb33-44d4-af94-7b24177ff046"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.091145 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5986e88e-5488-4faa-829f-eebfbf81f209" (UID: "5986e88e-5488-4faa-829f-eebfbf81f209"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.093799 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7" (OuterVolumeSpecName: "kube-api-access-tkpv7") pod "7ec53377-bb33-44d4-af94-7b24177ff046" (UID: "7ec53377-bb33-44d4-af94-7b24177ff046"). InnerVolumeSpecName "kube-api-access-tkpv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.095078 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp" (OuterVolumeSpecName: "kube-api-access-nh5mp") pod "c8faa97c-b754-433d-ae7c-d875a0cd286e" (UID: "c8faa97c-b754-433d-ae7c-d875a0cd286e"). InnerVolumeSpecName "kube-api-access-nh5mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.095843 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r" (OuterVolumeSpecName: "kube-api-access-gqk7r") pod "5986e88e-5488-4faa-829f-eebfbf81f209" (UID: "5986e88e-5488-4faa-829f-eebfbf81f209"). InnerVolumeSpecName "kube-api-access-gqk7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.133216 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.191831 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5986e88e-5488-4faa-829f-eebfbf81f209-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.191869 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkpv7\" (UniqueName: \"kubernetes.io/projected/7ec53377-bb33-44d4-af94-7b24177ff046-kube-api-access-tkpv7\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.191882 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqk7r\" (UniqueName: \"kubernetes.io/projected/5986e88e-5488-4faa-829f-eebfbf81f209-kube-api-access-gqk7r\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.191893 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ec53377-bb33-44d4-af94-7b24177ff046-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.191905 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh5mp\" (UniqueName: \"kubernetes.io/projected/c8faa97c-b754-433d-ae7c-d875a0cd286e-kube-api-access-nh5mp\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.319721 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"b09884bcfe897aaae2d5127126df71dbd8b6496dc990b7b03cfd3c2314f7431a"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.322572 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-29f2-account-create-update-qfrmt" event={"ID":"8dcfd4ee-cde3-49e8-ad50-af6fd23af517","Type":"ContainerDied","Data":"69cbe5b41f9ed22afdfe946e8e8e3017a70ebb2a0c87ee5d126d6ce2d1fe4c61"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.322596 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-29f2-account-create-update-qfrmt" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.322613 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69cbe5b41f9ed22afdfe946e8e8e3017a70ebb2a0c87ee5d126d6ce2d1fe4c61" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.325026 4867 generic.go:334] "Generic (PLEG): container finished" podID="a86271f9-3990-46b9-971d-0151ddda839f" containerID="e951181ca2240c8345911e4e0b3c224a258ab5437f10f82e6d47edbc589d9fb2" exitCode=0 Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.325088 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-b9p6s" event={"ID":"a86271f9-3990-46b9-971d-0151ddda839f","Type":"ContainerDied","Data":"e951181ca2240c8345911e4e0b3c224a258ab5437f10f82e6d47edbc589d9fb2"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.328011 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-74a1-account-create-update-mdf7g" event={"ID":"7ec53377-bb33-44d4-af94-7b24177ff046","Type":"ContainerDied","Data":"eb06b723ee7d265611fd63c4a317e53a4c6a41a51c16466b04d5110f87218818"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.328063 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb06b723ee7d265611fd63c4a317e53a4c6a41a51c16466b04d5110f87218818" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.328155 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-74a1-account-create-update-mdf7g" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.331155 4867 generic.go:334] "Generic (PLEG): container finished" podID="b9fabde1-97a0-4d14-8613-6bd031ab968d" containerID="77c9a26c17603a8e3b6947d790bbaaaac80e35891cbed056027c6b7726af2e06" exitCode=0 Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.331326 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dz67c" event={"ID":"b9fabde1-97a0-4d14-8613-6bd031ab968d","Type":"ContainerDied","Data":"77c9a26c17603a8e3b6947d790bbaaaac80e35891cbed056027c6b7726af2e06"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.332951 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b8be-account-create-update-29zkz" event={"ID":"5986e88e-5488-4faa-829f-eebfbf81f209","Type":"ContainerDied","Data":"25ab6bfd724a0c73d90b8c04e67e8060e8751b30650dd0ca96adcbb3f04d0bf8"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.332982 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25ab6bfd724a0c73d90b8c04e67e8060e8751b30650dd0ca96adcbb3f04d0bf8" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.333037 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b8be-account-create-update-29zkz" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.351856 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x8h7k" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.351844 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x8h7k" event={"ID":"c8faa97c-b754-433d-ae7c-d875a0cd286e","Type":"ContainerDied","Data":"cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37"} Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.351927 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb91f12523eb2d7e7b45568087aeda74d7c1617a7d4185f2c30570d9658bda37" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.676024 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.802694 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts\") pod \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.802875 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78ggr\" (UniqueName: \"kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr\") pod \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\" (UID: \"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51\") " Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.803610 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" (UID: "e6da4f5f-8f05-47d2-a3d5-79cfd7666d51"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.808512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr" (OuterVolumeSpecName: "kube-api-access-78ggr") pod "e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" (UID: "e6da4f5f-8f05-47d2-a3d5-79cfd7666d51"). InnerVolumeSpecName "kube-api-access-78ggr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.904852 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78ggr\" (UniqueName: \"kubernetes.io/projected/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-kube-api-access-78ggr\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:08 crc kubenswrapper[4867]: I1212 07:07:08.904889 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.370897 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qtct9" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.371879 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qtct9" event={"ID":"e6da4f5f-8f05-47d2-a3d5-79cfd7666d51","Type":"ContainerDied","Data":"437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3"} Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.371926 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="437e9a265698bb5c513d38426f08a2993a5eed9b162ec75c7151a9a56fc370c3" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.750704 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.760041 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822425 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822494 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822539 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822620 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqkf2\" (UniqueName: \"kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2\") pod \"b9fabde1-97a0-4d14-8613-6bd031ab968d\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822669 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvgtw\" (UniqueName: \"kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822725 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822773 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn\") pod \"a86271f9-3990-46b9-971d-0151ddda839f\" (UID: \"a86271f9-3990-46b9-971d-0151ddda839f\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822816 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts\") pod \"b9fabde1-97a0-4d14-8613-6bd031ab968d\" (UID: \"b9fabde1-97a0-4d14-8613-6bd031ab968d\") " Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822862 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run" (OuterVolumeSpecName: "var-run") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.822949 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823191 4867 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823204 4867 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823213 4867 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a86271f9-3990-46b9-971d-0151ddda839f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823414 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823419 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9fabde1-97a0-4d14-8613-6bd031ab968d" (UID: "b9fabde1-97a0-4d14-8613-6bd031ab968d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.823753 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts" (OuterVolumeSpecName: "scripts") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.826972 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2" (OuterVolumeSpecName: "kube-api-access-jqkf2") pod "b9fabde1-97a0-4d14-8613-6bd031ab968d" (UID: "b9fabde1-97a0-4d14-8613-6bd031ab968d"). InnerVolumeSpecName "kube-api-access-jqkf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.827730 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw" (OuterVolumeSpecName: "kube-api-access-qvgtw") pod "a86271f9-3990-46b9-971d-0151ddda839f" (UID: "a86271f9-3990-46b9-971d-0151ddda839f"). InnerVolumeSpecName "kube-api-access-qvgtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.925291 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvgtw\" (UniqueName: \"kubernetes.io/projected/a86271f9-3990-46b9-971d-0151ddda839f-kube-api-access-qvgtw\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.925337 4867 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.925351 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9fabde1-97a0-4d14-8613-6bd031ab968d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.925367 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a86271f9-3990-46b9-971d-0151ddda839f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:09 crc kubenswrapper[4867]: I1212 07:07:09.925381 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqkf2\" (UniqueName: \"kubernetes.io/projected/b9fabde1-97a0-4d14-8613-6bd031ab968d-kube-api-access-jqkf2\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.402201 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-b9p6s" event={"ID":"a86271f9-3990-46b9-971d-0151ddda839f","Type":"ContainerDied","Data":"76fd69aa3c2d1878f209b6a29c4102f2c3ad87c59c62eb8795673bc33e0dc2e5"} Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.402267 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76fd69aa3c2d1878f209b6a29c4102f2c3ad87c59c62eb8795673bc33e0dc2e5" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.402219 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-b9p6s" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.403647 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dz67c" event={"ID":"b9fabde1-97a0-4d14-8613-6bd031ab968d","Type":"ContainerDied","Data":"85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58"} Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.403960 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85556c586e2b2db97ff22579bb72c31f3e19441315d62d9b47e9af349d9e7c58" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.404015 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dz67c" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.456061 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8njtj-config-b9p6s"] Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.464985 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8njtj-config-b9p6s"] Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556252 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8njtj-config-flbcx"] Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556620 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcfd4ee-cde3-49e8-ad50-af6fd23af517" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556636 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcfd4ee-cde3-49e8-ad50-af6fd23af517" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556652 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ec53377-bb33-44d4-af94-7b24177ff046" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556658 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ec53377-bb33-44d4-af94-7b24177ff046" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556675 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5986e88e-5488-4faa-829f-eebfbf81f209" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556681 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5986e88e-5488-4faa-829f-eebfbf81f209" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556691 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8faa97c-b754-433d-ae7c-d875a0cd286e" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556697 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8faa97c-b754-433d-ae7c-d875a0cd286e" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556704 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1432b4ca-14f2-4044-a58c-d93ff9930e17" containerName="swift-ring-rebalance" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556710 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1432b4ca-14f2-4044-a58c-d93ff9930e17" containerName="swift-ring-rebalance" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556716 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556722 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556730 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fabde1-97a0-4d14-8613-6bd031ab968d" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556735 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fabde1-97a0-4d14-8613-6bd031ab968d" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: E1212 07:07:10.556748 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86271f9-3990-46b9-971d-0151ddda839f" containerName="ovn-config" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556755 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86271f9-3990-46b9-971d-0151ddda839f" containerName="ovn-config" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556915 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556927 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcfd4ee-cde3-49e8-ad50-af6fd23af517" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556941 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5986e88e-5488-4faa-829f-eebfbf81f209" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556950 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ec53377-bb33-44d4-af94-7b24177ff046" containerName="mariadb-account-create-update" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556963 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1432b4ca-14f2-4044-a58c-d93ff9930e17" containerName="swift-ring-rebalance" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556971 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9fabde1-97a0-4d14-8613-6bd031ab968d" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556978 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8faa97c-b754-433d-ae7c-d875a0cd286e" containerName="mariadb-database-create" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.556989 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86271f9-3990-46b9-971d-0151ddda839f" containerName="ovn-config" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.557587 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.559826 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.574435 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj-config-flbcx"] Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636106 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636166 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636194 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636292 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636323 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lrbf\" (UniqueName: \"kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.636357 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737563 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737616 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737635 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737673 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737690 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lrbf\" (UniqueName: \"kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.737718 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.738779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.738782 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.739191 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.739964 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.740430 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.764982 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lrbf\" (UniqueName: \"kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf\") pod \"ovn-controller-8njtj-config-flbcx\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.850548 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86271f9-3990-46b9-971d-0151ddda839f" path="/var/lib/kubelet/pods/a86271f9-3990-46b9-971d-0151ddda839f/volumes" Dec 12 07:07:10 crc kubenswrapper[4867]: I1212 07:07:10.877301 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:11 crc kubenswrapper[4867]: I1212 07:07:11.409741 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8njtj-config-flbcx"] Dec 12 07:07:14 crc kubenswrapper[4867]: W1212 07:07:14.195187 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f779e95_be7c_43f9_900e_513c76d60ba1.slice/crio-81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1 WatchSource:0}: Error finding container 81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1: Status 404 returned error can't find the container with id 81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1 Dec 12 07:07:14 crc kubenswrapper[4867]: I1212 07:07:14.448121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-flbcx" event={"ID":"0f779e95-be7c-43f9-900e-513c76d60ba1","Type":"ContainerStarted","Data":"81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1"} Dec 12 07:07:15 crc kubenswrapper[4867]: I1212 07:07:15.460107 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n2xp2" event={"ID":"d930b18c-8189-43e2-b59b-9b75e0ac338b","Type":"ContainerStarted","Data":"fe84ae831036079376416809cc68115d051b2918c176b712ed73dbcf926065ad"} Dec 12 07:07:15 crc kubenswrapper[4867]: I1212 07:07:15.464003 4867 generic.go:334] "Generic (PLEG): container finished" podID="0f779e95-be7c-43f9-900e-513c76d60ba1" containerID="4577bfb6bb31cfe2a93939f5385f788d8aa1c3bb19025613cf01defc00d6a24d" exitCode=0 Dec 12 07:07:15 crc kubenswrapper[4867]: I1212 07:07:15.464047 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-flbcx" event={"ID":"0f779e95-be7c-43f9-900e-513c76d60ba1","Type":"ContainerDied","Data":"4577bfb6bb31cfe2a93939f5385f788d8aa1c3bb19025613cf01defc00d6a24d"} Dec 12 07:07:15 crc kubenswrapper[4867]: I1212 07:07:15.484380 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-n2xp2" podStartSLOduration=5.597564803 podStartE2EDuration="14.484359064s" podCreationTimestamp="2025-12-12 07:07:01 +0000 UTC" firstStartedPulling="2025-12-12 07:07:05.485573439 +0000 UTC m=+1113.056954708" lastFinishedPulling="2025-12-12 07:07:14.3723677 +0000 UTC m=+1121.943748969" observedRunningTime="2025-12-12 07:07:15.476336156 +0000 UTC m=+1123.047717435" watchObservedRunningTime="2025-12-12 07:07:15.484359064 +0000 UTC m=+1123.055740333" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.474003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb"} Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.474379 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab"} Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.789774 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861581 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861646 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861703 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861767 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861829 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.861865 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lrbf\" (UniqueName: \"kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf\") pod \"0f779e95-be7c-43f9-900e-513c76d60ba1\" (UID: \"0f779e95-be7c-43f9-900e-513c76d60ba1\") " Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862128 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862157 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862159 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run" (OuterVolumeSpecName: "var-run") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862455 4867 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862474 4867 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.862485 4867 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0f779e95-be7c-43f9-900e-513c76d60ba1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.865118 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.865660 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts" (OuterVolumeSpecName: "scripts") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.868391 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf" (OuterVolumeSpecName: "kube-api-access-2lrbf") pod "0f779e95-be7c-43f9-900e-513c76d60ba1" (UID: "0f779e95-be7c-43f9-900e-513c76d60ba1"). InnerVolumeSpecName "kube-api-access-2lrbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.964672 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lrbf\" (UniqueName: \"kubernetes.io/projected/0f779e95-be7c-43f9-900e-513c76d60ba1-kube-api-access-2lrbf\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.964711 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:16 crc kubenswrapper[4867]: I1212 07:07:16.964723 4867 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0f779e95-be7c-43f9-900e-513c76d60ba1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.489411 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj-config-flbcx" event={"ID":"0f779e95-be7c-43f9-900e-513c76d60ba1","Type":"ContainerDied","Data":"81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1"} Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.489840 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81e4720df2ddb95a80b160adb6383263e113ab45021db7d5c329446eb051fde1" Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.489928 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj-config-flbcx" Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.494003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751"} Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.494059 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc"} Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.872332 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8njtj-config-flbcx"] Dec 12 07:07:17 crc kubenswrapper[4867]: I1212 07:07:17.882038 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8njtj-config-flbcx"] Dec 12 07:07:18 crc kubenswrapper[4867]: I1212 07:07:18.848950 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f779e95-be7c-43f9-900e-513c76d60ba1" path="/var/lib/kubelet/pods/0f779e95-be7c-43f9-900e-513c76d60ba1/volumes" Dec 12 07:07:24 crc kubenswrapper[4867]: I1212 07:07:24.559763 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178"} Dec 12 07:07:24 crc kubenswrapper[4867]: I1212 07:07:24.562334 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7lxjs" event={"ID":"a3a5237b-ffff-43b1-a921-68affabec6cd","Type":"ContainerStarted","Data":"bcf8eb7e4e5ab4fec8f5c52187980849509ed93d855f00433e7a743ea90ffb05"} Dec 12 07:07:24 crc kubenswrapper[4867]: I1212 07:07:24.586933 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-7lxjs" podStartSLOduration=2.367041554 podStartE2EDuration="37.586916553s" podCreationTimestamp="2025-12-12 07:06:47 +0000 UTC" firstStartedPulling="2025-12-12 07:06:48.446277368 +0000 UTC m=+1096.017658637" lastFinishedPulling="2025-12-12 07:07:23.666152367 +0000 UTC m=+1131.237533636" observedRunningTime="2025-12-12 07:07:24.578786833 +0000 UTC m=+1132.150168102" watchObservedRunningTime="2025-12-12 07:07:24.586916553 +0000 UTC m=+1132.158297822" Dec 12 07:07:25 crc kubenswrapper[4867]: I1212 07:07:25.573961 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54"} Dec 12 07:07:25 crc kubenswrapper[4867]: I1212 07:07:25.574360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec"} Dec 12 07:07:25 crc kubenswrapper[4867]: I1212 07:07:25.574375 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd"} Dec 12 07:07:29 crc kubenswrapper[4867]: I1212 07:07:29.616098 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d"} Dec 12 07:07:29 crc kubenswrapper[4867]: I1212 07:07:29.616676 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d"} Dec 12 07:07:30 crc kubenswrapper[4867]: I1212 07:07:30.629012 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070"} Dec 12 07:07:30 crc kubenswrapper[4867]: I1212 07:07:30.629311 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c"} Dec 12 07:07:30 crc kubenswrapper[4867]: I1212 07:07:30.629324 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0"} Dec 12 07:07:30 crc kubenswrapper[4867]: I1212 07:07:30.629333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309"} Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.645872 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerStarted","Data":"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb"} Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.697835 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.71191461 podStartE2EDuration="57.697811992s" podCreationTimestamp="2025-12-12 07:06:34 +0000 UTC" firstStartedPulling="2025-12-12 07:07:08.139954407 +0000 UTC m=+1115.711335686" lastFinishedPulling="2025-12-12 07:07:29.125851799 +0000 UTC m=+1136.697233068" observedRunningTime="2025-12-12 07:07:31.684570688 +0000 UTC m=+1139.255951967" watchObservedRunningTime="2025-12-12 07:07:31.697811992 +0000 UTC m=+1139.269193261" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.965159 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:07:31 crc kubenswrapper[4867]: E1212 07:07:31.966079 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f779e95-be7c-43f9-900e-513c76d60ba1" containerName="ovn-config" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.966135 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f779e95-be7c-43f9-900e-513c76d60ba1" containerName="ovn-config" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.966670 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f779e95-be7c-43f9-900e-513c76d60ba1" containerName="ovn-config" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.968260 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.970823 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 12 07:07:31 crc kubenswrapper[4867]: I1212 07:07:31.978313 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011214 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011335 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjgs\" (UniqueName: \"kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011366 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011493 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.011535 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113062 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113099 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjgs\" (UniqueName: \"kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113120 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113176 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113214 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.113268 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.114344 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.114368 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.114522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.114625 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.115213 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.145008 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjgs\" (UniqueName: \"kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs\") pod \"dnsmasq-dns-8db84466c-9tz5f\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.287848 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.666936 4867 generic.go:334] "Generic (PLEG): container finished" podID="d930b18c-8189-43e2-b59b-9b75e0ac338b" containerID="fe84ae831036079376416809cc68115d051b2918c176b712ed73dbcf926065ad" exitCode=0 Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.667077 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n2xp2" event={"ID":"d930b18c-8189-43e2-b59b-9b75e0ac338b","Type":"ContainerDied","Data":"fe84ae831036079376416809cc68115d051b2918c176b712ed73dbcf926065ad"} Dec 12 07:07:32 crc kubenswrapper[4867]: I1212 07:07:32.786733 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:07:33 crc kubenswrapper[4867]: I1212 07:07:33.677606 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerStarted","Data":"9093b22e58464b9a68507926c5e4eef2f36c414ba9a29175fac943e3c3501e62"} Dec 12 07:07:33 crc kubenswrapper[4867]: I1212 07:07:33.677917 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerStarted","Data":"7f3f05b8058bb51f6686820e54789ed0b54cbe51b9583abd0e4a16b408869122"} Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.060285 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.150415 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data\") pod \"d930b18c-8189-43e2-b59b-9b75e0ac338b\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.150501 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmxhr\" (UniqueName: \"kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr\") pod \"d930b18c-8189-43e2-b59b-9b75e0ac338b\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.150596 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle\") pod \"d930b18c-8189-43e2-b59b-9b75e0ac338b\" (UID: \"d930b18c-8189-43e2-b59b-9b75e0ac338b\") " Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.157592 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr" (OuterVolumeSpecName: "kube-api-access-wmxhr") pod "d930b18c-8189-43e2-b59b-9b75e0ac338b" (UID: "d930b18c-8189-43e2-b59b-9b75e0ac338b"). InnerVolumeSpecName "kube-api-access-wmxhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.176392 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d930b18c-8189-43e2-b59b-9b75e0ac338b" (UID: "d930b18c-8189-43e2-b59b-9b75e0ac338b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.204926 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data" (OuterVolumeSpecName: "config-data") pod "d930b18c-8189-43e2-b59b-9b75e0ac338b" (UID: "d930b18c-8189-43e2-b59b-9b75e0ac338b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.252107 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.252143 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmxhr\" (UniqueName: \"kubernetes.io/projected/d930b18c-8189-43e2-b59b-9b75e0ac338b-kube-api-access-wmxhr\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.252154 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d930b18c-8189-43e2-b59b-9b75e0ac338b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.687106 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n2xp2" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.687120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n2xp2" event={"ID":"d930b18c-8189-43e2-b59b-9b75e0ac338b","Type":"ContainerDied","Data":"644ce48e06096cb81c11ec4f19ce0e17dbdce9809b491e58d8e8df0f0967120d"} Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.687567 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="644ce48e06096cb81c11ec4f19ce0e17dbdce9809b491e58d8e8df0f0967120d" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.689547 4867 generic.go:334] "Generic (PLEG): container finished" podID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerID="9093b22e58464b9a68507926c5e4eef2f36c414ba9a29175fac943e3c3501e62" exitCode=0 Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.689606 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerDied","Data":"9093b22e58464b9a68507926c5e4eef2f36c414ba9a29175fac943e3c3501e62"} Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.941298 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.958915 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-92kwf"] Dec 12 07:07:34 crc kubenswrapper[4867]: E1212 07:07:34.966602 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d930b18c-8189-43e2-b59b-9b75e0ac338b" containerName="keystone-db-sync" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.966642 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d930b18c-8189-43e2-b59b-9b75e0ac338b" containerName="keystone-db-sync" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.966858 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d930b18c-8189-43e2-b59b-9b75e0ac338b" containerName="keystone-db-sync" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.967501 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.971323 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.971471 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.971538 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.972303 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9jg9" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.972588 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 07:07:34 crc kubenswrapper[4867]: I1212 07:07:34.981738 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-92kwf"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.016920 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.025147 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.059454 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.154587 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-ptmqx"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.156328 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.160147 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.160273 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-k65fc" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.161896 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.166860 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.166909 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7mf9\" (UniqueName: \"kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.166935 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.166961 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.166978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqcd4\" (UniqueName: \"kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.167682 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.167953 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.168014 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.168096 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.168163 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.168273 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.168314 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.200390 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7fnd8"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.202686 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.214450 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.214816 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.221834 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-r2vqq" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.229454 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ptmqx"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.257957 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7fnd8"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.270532 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.270834 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.270969 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.271121 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67j8q\" (UniqueName: \"kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.271266 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.271916 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272062 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272179 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272327 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42v2r\" (UniqueName: \"kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272444 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272567 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272795 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7mf9\" (UniqueName: \"kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.272921 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273040 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273156 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273288 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqcd4\" (UniqueName: \"kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273437 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273667 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273767 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.273871 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.275428 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.276265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.276391 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.282772 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.289986 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.291010 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.291703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.292788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.297926 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqcd4\" (UniqueName: \"kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.299550 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7mf9\" (UniqueName: \"kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9\") pod \"dnsmasq-dns-767d96458c-2mxj9\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.311643 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9dbsw"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.313921 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.316372 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys\") pod \"keystone-bootstrap-92kwf\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.317698 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qqxfz" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.319032 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.338347 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9dbsw"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.361381 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.366002 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.374061 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376386 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67j8q\" (UniqueName: \"kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376443 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmvlb\" (UniqueName: \"kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376474 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376517 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42v2r\" (UniqueName: \"kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376543 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376573 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376611 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqc6f\" (UniqueName: \"kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376624 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376652 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376699 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376721 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376743 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376822 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376844 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376850 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376875 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376933 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.376978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.377028 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.377056 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.383429 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.385707 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.386058 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.392798 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.395559 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.397666 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.410121 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.419182 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.421102 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42v2r\" (UniqueName: \"kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r\") pod \"cinder-db-sync-7fnd8\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.460507 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67j8q\" (UniqueName: \"kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q\") pod \"neutron-db-sync-ptmqx\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.462128 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6lmsp"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.463583 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.472475 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6lmsp"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.473601 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.473987 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-hqhp8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.474417 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.490370 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmvlb\" (UniqueName: \"kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.490666 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.490799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.490881 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqc6f\" (UniqueName: \"kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.490966 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491289 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491376 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491475 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491581 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491688 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491884 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.491965 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pck7x\" (UniqueName: \"kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.492039 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.492182 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.500123 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.503741 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.505503 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.509808 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.510887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.514930 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.515074 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.535674 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.540564 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmvlb\" (UniqueName: \"kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.550806 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqc6f\" (UniqueName: \"kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f\") pod \"barbican-db-sync-9dbsw\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.556441 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.568981 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.572510 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.572547 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts\") pod \"ceilometer-0\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.573953 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.581580 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.582395 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.582855 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.588888 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.596988 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdnzk\" (UniqueName: \"kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.604400 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pck7x\" (UniqueName: \"kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.604461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.604579 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.604635 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.604720 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605293 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605477 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605521 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605553 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605602 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.605721 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.620599 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.623726 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.624444 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.630089 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pck7x\" (UniqueName: \"kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x\") pod \"placement-db-sync-6lmsp\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708583 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708632 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708771 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdnzk\" (UniqueName: \"kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708829 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708854 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.708882 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.711009 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.711884 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.711022 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.712656 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.726905 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.737081 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdnzk\" (UniqueName: \"kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk\") pod \"dnsmasq-dns-7fc6d4ffc7-8jtzn\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.896477 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lmsp" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.932760 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:35 crc kubenswrapper[4867]: I1212 07:07:35.999316 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:07:36 crc kubenswrapper[4867]: W1212 07:07:36.004337 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdcab93e_7564_4d7e_b4fb_2da47c638c4e.slice/crio-55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534 WatchSource:0}: Error finding container 55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534: Status 404 returned error can't find the container with id 55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534 Dec 12 07:07:36 crc kubenswrapper[4867]: I1212 07:07:36.181091 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9dbsw"] Dec 12 07:07:36 crc kubenswrapper[4867]: W1212 07:07:36.195411 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68801d28_d85f_46b6_86cd_debaf6dab7ca.slice/crio-19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784 WatchSource:0}: Error finding container 19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784: Status 404 returned error can't find the container with id 19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784 Dec 12 07:07:36 crc kubenswrapper[4867]: I1212 07:07:36.283703 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ptmqx"] Dec 12 07:07:36 crc kubenswrapper[4867]: I1212 07:07:36.295467 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7fnd8"] Dec 12 07:07:36 crc kubenswrapper[4867]: W1212 07:07:36.300698 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f78a88a_c256_4e35_8f64_1c155650cca6.slice/crio-dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235 WatchSource:0}: Error finding container dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235: Status 404 returned error can't find the container with id dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235 Dec 12 07:07:36 crc kubenswrapper[4867]: W1212 07:07:36.315297 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63341abf_5801_4239_8370_4c0732c95e3b.slice/crio-c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1 WatchSource:0}: Error finding container c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1: Status 404 returned error can't find the container with id c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1 Dec 12 07:07:36 crc kubenswrapper[4867]: I1212 07:07:36.416143 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-92kwf"] Dec 12 07:07:36 crc kubenswrapper[4867]: W1212 07:07:36.424523 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd501d207_060c_4e80_b631_bfe7a6754c05.slice/crio-ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8 WatchSource:0}: Error finding container ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8: Status 404 returned error can't find the container with id ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8 Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.475676 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.566059 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.588209 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6lmsp"] Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.737456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" event={"ID":"cdcab93e-7564-4d7e-b4fb-2da47c638c4e","Type":"ContainerStarted","Data":"55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.739107 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" event={"ID":"bf3306f4-eec2-4188-b964-ac1cec81824a","Type":"ContainerStarted","Data":"dd837fc428c888b284c80683229a242eb4010f3b4aa4a44f78117cb707ad05c7"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.740101 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92kwf" event={"ID":"d501d207-060c-4e80-b631-bfe7a6754c05","Type":"ContainerStarted","Data":"ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.741118 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lmsp" event={"ID":"7299cd15-ce08-48f3-be09-c3a70901c4aa","Type":"ContainerStarted","Data":"07e77352529ae62199afdaaa6bce28b46840142d0f0e38693a25ada31e6a71df"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.742214 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerStarted","Data":"05018dcd51b85f6cc97091f9c80d414dcf78b114274060274ccde5a89ff26c49"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.743404 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7fnd8" event={"ID":"63341abf-5801-4239-8370-4c0732c95e3b","Type":"ContainerStarted","Data":"c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.744567 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ptmqx" event={"ID":"3f78a88a-c256-4e35-8f64-1c155650cca6","Type":"ContainerStarted","Data":"dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.745452 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9dbsw" event={"ID":"68801d28-d85f-46b6-86cd-debaf6dab7ca","Type":"ContainerStarted","Data":"19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.747206 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerStarted","Data":"63bf09d8ac7be5ac381adb2e7a7413967e43e8b8f5634ed64f16a2bfcf37a4d4"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.747330 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" containerID="cri-o://63bf09d8ac7be5ac381adb2e7a7413967e43e8b8f5634ed64f16a2bfcf37a4d4" gracePeriod=10 Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:36.747408 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.667494 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podStartSLOduration=6.667471091 podStartE2EDuration="6.667471091s" podCreationTimestamp="2025-12-12 07:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:36.769825433 +0000 UTC m=+1144.341206702" watchObservedRunningTime="2025-12-12 07:07:37.667471091 +0000 UTC m=+1145.238852360" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.670425 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.774207 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92kwf" event={"ID":"d501d207-060c-4e80-b631-bfe7a6754c05","Type":"ContainerStarted","Data":"ebaea2e9bb46bdb9b9f5d3e785667e43809d4ca7be867bac6014e24edd424184"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.777575 4867 generic.go:334] "Generic (PLEG): container finished" podID="cdcab93e-7564-4d7e-b4fb-2da47c638c4e" containerID="8e0865343bc535ead2fc7e07bf96597f4556a3b0218b2ab297f2420213bffa0f" exitCode=0 Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.777634 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" event={"ID":"cdcab93e-7564-4d7e-b4fb-2da47c638c4e","Type":"ContainerDied","Data":"8e0865343bc535ead2fc7e07bf96597f4556a3b0218b2ab297f2420213bffa0f"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.786287 4867 generic.go:334] "Generic (PLEG): container finished" podID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerID="63bf09d8ac7be5ac381adb2e7a7413967e43e8b8f5634ed64f16a2bfcf37a4d4" exitCode=0 Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.786368 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerDied","Data":"63bf09d8ac7be5ac381adb2e7a7413967e43e8b8f5634ed64f16a2bfcf37a4d4"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.791312 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ptmqx" event={"ID":"3f78a88a-c256-4e35-8f64-1c155650cca6","Type":"ContainerStarted","Data":"77646526d618a77ee8b6ba9c19111bdb9e0e386abe004e3170ce8918fdae8267"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.803641 4867 generic.go:334] "Generic (PLEG): container finished" podID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerID="11d7633344c9a96316d54f1335579e380441baa68368ba0bd6c2d213c800734d" exitCode=0 Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.803691 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" event={"ID":"bf3306f4-eec2-4188-b964-ac1cec81824a","Type":"ContainerDied","Data":"11d7633344c9a96316d54f1335579e380441baa68368ba0bd6c2d213c800734d"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.804512 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-92kwf" podStartSLOduration=3.8044888869999998 podStartE2EDuration="3.804488887s" podCreationTimestamp="2025-12-12 07:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:37.796113141 +0000 UTC m=+1145.367494430" watchObservedRunningTime="2025-12-12 07:07:37.804488887 +0000 UTC m=+1145.375870146" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:37.904589 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-ptmqx" podStartSLOduration=2.904559174 podStartE2EDuration="2.904559174s" podCreationTimestamp="2025-12-12 07:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:37.860430051 +0000 UTC m=+1145.431811320" watchObservedRunningTime="2025-12-12 07:07:37.904559174 +0000 UTC m=+1145.475940443" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:41.866915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" event={"ID":"bf3306f4-eec2-4188-b964-ac1cec81824a","Type":"ContainerStarted","Data":"7046e803bfbfbbe7b0a68f19ce0ff19c35b971902172ac64fe4af0e5a0aff4bc"} Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:41.867672 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:41.893806 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" podStartSLOduration=6.89378278 podStartE2EDuration="6.89378278s" podCreationTimestamp="2025-12-12 07:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:07:41.887877104 +0000 UTC m=+1149.459258383" watchObservedRunningTime="2025-12-12 07:07:41.89378278 +0000 UTC m=+1149.465164049" Dec 12 07:07:42 crc kubenswrapper[4867]: I1212 07:07:42.296844 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Dec 12 07:07:45 crc kubenswrapper[4867]: I1212 07:07:45.934986 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:07:46 crc kubenswrapper[4867]: I1212 07:07:46.004333 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:07:46 crc kubenswrapper[4867]: I1212 07:07:46.004570 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" containerID="cri-o://6cbb9d6eb9e947faca9544cd025ec52e5ab4eabca3e78cfb9f4d07e4d9f9b9b4" gracePeriod=10 Dec 12 07:07:49 crc kubenswrapper[4867]: I1212 07:07:49.353759 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Dec 12 07:07:52 crc kubenswrapper[4867]: I1212 07:07:52.294168 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 12 07:07:52 crc kubenswrapper[4867]: I1212 07:07:52.969847 4867 generic.go:334] "Generic (PLEG): container finished" podID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerID="6cbb9d6eb9e947faca9544cd025ec52e5ab4eabca3e78cfb9f4d07e4d9f9b9b4" exitCode=0 Dec 12 07:07:52 crc kubenswrapper[4867]: I1212 07:07:52.969882 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" event={"ID":"b0f75dd4-02a1-473c-a06f-c8477baa2d90","Type":"ContainerDied","Data":"6cbb9d6eb9e947faca9544cd025ec52e5ab4eabca3e78cfb9f4d07e4d9f9b9b4"} Dec 12 07:07:53 crc kubenswrapper[4867]: E1212 07:07:53.236850 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16" Dec 12 07:07:53 crc kubenswrapper[4867]: E1212 07:07:53.237089 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqc6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9dbsw_openstack(68801d28-d85f-46b6-86cd-debaf6dab7ca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:07:53 crc kubenswrapper[4867]: E1212 07:07:53.238301 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9dbsw" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" Dec 12 07:07:53 crc kubenswrapper[4867]: E1212 07:07:53.982762 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16\\\"\"" pod="openstack/barbican-db-sync-9dbsw" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" Dec 12 07:07:57 crc kubenswrapper[4867]: I1212 07:07:57.295496 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 12 07:07:59 crc kubenswrapper[4867]: I1212 07:07:59.353275 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.034916 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" event={"ID":"cdcab93e-7564-4d7e-b4fb-2da47c638c4e","Type":"ContainerDied","Data":"55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534"} Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.034955 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55de591210277b49bb5b111f9ed9693e38d8903ed27b98f9d53ada68d5bc0534" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.093924 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204360 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204500 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204605 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204688 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7mf9\" (UniqueName: \"kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204743 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.204776 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb\") pod \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\" (UID: \"cdcab93e-7564-4d7e-b4fb-2da47c638c4e\") " Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.210545 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9" (OuterVolumeSpecName: "kube-api-access-k7mf9") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "kube-api-access-k7mf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.228285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.229906 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.232256 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.233119 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config" (OuterVolumeSpecName: "config") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.249547 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cdcab93e-7564-4d7e-b4fb-2da47c638c4e" (UID: "cdcab93e-7564-4d7e-b4fb-2da47c638c4e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307314 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307345 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307357 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307366 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307375 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:01 crc kubenswrapper[4867]: I1212 07:08:01.307383 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7mf9\" (UniqueName: \"kubernetes.io/projected/cdcab93e-7564-4d7e-b4fb-2da47c638c4e-kube-api-access-k7mf9\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:02 crc kubenswrapper[4867]: I1212 07:08:02.045529 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-767d96458c-2mxj9" Dec 12 07:08:02 crc kubenswrapper[4867]: I1212 07:08:02.117555 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:08:02 crc kubenswrapper[4867]: I1212 07:08:02.133040 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-767d96458c-2mxj9"] Dec 12 07:08:02 crc kubenswrapper[4867]: I1212 07:08:02.296874 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 12 07:08:02 crc kubenswrapper[4867]: I1212 07:08:02.850769 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdcab93e-7564-4d7e-b4fb-2da47c638c4e" path="/var/lib/kubelet/pods/cdcab93e-7564-4d7e-b4fb-2da47c638c4e/volumes" Dec 12 07:08:03 crc kubenswrapper[4867]: I1212 07:08:03.055343 4867 generic.go:334] "Generic (PLEG): container finished" podID="d501d207-060c-4e80-b631-bfe7a6754c05" containerID="ebaea2e9bb46bdb9b9f5d3e785667e43809d4ca7be867bac6014e24edd424184" exitCode=0 Dec 12 07:08:03 crc kubenswrapper[4867]: I1212 07:08:03.055386 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92kwf" event={"ID":"d501d207-060c-4e80-b631-bfe7a6754c05","Type":"ContainerDied","Data":"ebaea2e9bb46bdb9b9f5d3e785667e43809d4ca7be867bac6014e24edd424184"} Dec 12 07:08:04 crc kubenswrapper[4867]: I1212 07:08:04.354025 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 12 07:08:04 crc kubenswrapper[4867]: I1212 07:08:04.354760 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:08:07 crc kubenswrapper[4867]: I1212 07:08:07.297928 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 12 07:08:09 crc kubenswrapper[4867]: I1212 07:08:09.354875 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 12 07:08:10 crc kubenswrapper[4867]: E1212 07:08:10.413262 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49" Dec 12 07:08:10 crc kubenswrapper[4867]: E1212 07:08:10.413471 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-42v2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7fnd8_openstack(63341abf-5801-4239-8370-4c0732c95e3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:08:10 crc kubenswrapper[4867]: E1212 07:08:10.414728 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7fnd8" podUID="63341abf-5801-4239-8370-4c0732c95e3b" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.472808 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.484498 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573166 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config\") pod \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573609 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573642 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573709 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb\") pod \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573761 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb\") pod \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573787 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqcd4\" (UniqueName: \"kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573816 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc\") pod \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573839 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573885 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573916 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys\") pod \"d501d207-060c-4e80-b631-bfe7a6754c05\" (UID: \"d501d207-060c-4e80-b631-bfe7a6754c05\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.573961 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trbc5\" (UniqueName: \"kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5\") pod \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\" (UID: \"b0f75dd4-02a1-473c-a06f-c8477baa2d90\") " Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.579751 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5" (OuterVolumeSpecName: "kube-api-access-trbc5") pod "b0f75dd4-02a1-473c-a06f-c8477baa2d90" (UID: "b0f75dd4-02a1-473c-a06f-c8477baa2d90"). InnerVolumeSpecName "kube-api-access-trbc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.580649 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts" (OuterVolumeSpecName: "scripts") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.583618 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.594646 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.621515 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4" (OuterVolumeSpecName: "kube-api-access-pqcd4") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "kube-api-access-pqcd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.627045 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data" (OuterVolumeSpecName: "config-data") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.654489 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d501d207-060c-4e80-b631-bfe7a6754c05" (UID: "d501d207-060c-4e80-b631-bfe7a6754c05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.659391 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0f75dd4-02a1-473c-a06f-c8477baa2d90" (UID: "b0f75dd4-02a1-473c-a06f-c8477baa2d90"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.664362 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config" (OuterVolumeSpecName: "config") pod "b0f75dd4-02a1-473c-a06f-c8477baa2d90" (UID: "b0f75dd4-02a1-473c-a06f-c8477baa2d90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.665123 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0f75dd4-02a1-473c-a06f-c8477baa2d90" (UID: "b0f75dd4-02a1-473c-a06f-c8477baa2d90"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.672890 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0f75dd4-02a1-473c-a06f-c8477baa2d90" (UID: "b0f75dd4-02a1-473c-a06f-c8477baa2d90"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676523 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676554 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqcd4\" (UniqueName: \"kubernetes.io/projected/d501d207-060c-4e80-b631-bfe7a6754c05-kube-api-access-pqcd4\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676567 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676579 4867 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676590 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676600 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676608 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trbc5\" (UniqueName: \"kubernetes.io/projected/b0f75dd4-02a1-473c-a06f-c8477baa2d90-kube-api-access-trbc5\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676618 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676628 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676636 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d501d207-060c-4e80-b631-bfe7a6754c05-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:10 crc kubenswrapper[4867]: I1212 07:08:10.676643 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0f75dd4-02a1-473c-a06f-c8477baa2d90-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.116839 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" event={"ID":"b0f75dd4-02a1-473c-a06f-c8477baa2d90","Type":"ContainerDied","Data":"651ea1cdb6a0fe7d193b7083c2f24a8ef8f5dde9eb84a51e11a08099ba9f3f8d"} Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.116872 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.116891 4867 scope.go:117] "RemoveContainer" containerID="6cbb9d6eb9e947faca9544cd025ec52e5ab4eabca3e78cfb9f4d07e4d9f9b9b4" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.119432 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-92kwf" event={"ID":"d501d207-060c-4e80-b631-bfe7a6754c05","Type":"ContainerDied","Data":"ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8"} Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.119589 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe769c22bdd493c45b8b37e374e61c219a9ca01dff1f1370fe3a07b4306b7c8" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.119524 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-92kwf" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.121021 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49\\\"\"" pod="openstack/cinder-db-sync-7fnd8" podUID="63341abf-5801-4239-8370-4c0732c95e3b" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.155414 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.161606 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67fdf7998c-rlltl"] Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.163550 4867 scope.go:117] "RemoveContainer" containerID="942a76f37a7a5e61ee395a3c4846d0cf9ed5d2e8446b251c7f32e5379bd97c85" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.175141 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.229012 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5a548c25fe3d02f7a042cb0a6d28fc8039a34c4a3b3d07aadda4aba3a926e777" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.229214 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:5a548c25fe3d02f7a042cb0a6d28fc8039a34c4a3b3d07aadda4aba3a926e777,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n84hdchf5h5dfhch77h554h5cdh644h59bh57h58h597h5c8h55bh697h5fbh59h97hd8h64bh666h68bh5d6h676h699h555h79h78h65ch67fh5d7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmvlb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(de5ba6ce-674c-4887-8ba1-033e92e402c6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285403 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285460 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285512 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285694 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285737 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.285792 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnjgs\" (UniqueName: \"kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs\") pod \"94510a33-be0c-4f99-8fee-ad22a1b86719\" (UID: \"94510a33-be0c-4f99-8fee-ad22a1b86719\") " Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.291837 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs" (OuterVolumeSpecName: "kube-api-access-lnjgs") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "kube-api-access-lnjgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.330578 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.331633 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.335704 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.343356 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.343802 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config" (OuterVolumeSpecName: "config") pod "94510a33-be0c-4f99-8fee-ad22a1b86719" (UID: "94510a33-be0c-4f99-8fee-ad22a1b86719"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.387749 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.388054 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.388150 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnjgs\" (UniqueName: \"kubernetes.io/projected/94510a33-be0c-4f99-8fee-ad22a1b86719-kube-api-access-lnjgs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.388913 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.389002 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.389273 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/94510a33-be0c-4f99-8fee-ad22a1b86719-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.568261 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-92kwf"] Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.575560 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-92kwf"] Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.670495 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b5bbr"] Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.670914 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.670925 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.670939 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.670946 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.670958 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.670964 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.670977 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.670983 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.670997 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdcab93e-7564-4d7e-b4fb-2da47c638c4e" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671002 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdcab93e-7564-4d7e-b4fb-2da47c638c4e" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: E1212 07:08:11.671095 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d501d207-060c-4e80-b631-bfe7a6754c05" containerName="keystone-bootstrap" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671103 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d501d207-060c-4e80-b631-bfe7a6754c05" containerName="keystone-bootstrap" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671293 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d501d207-060c-4e80-b631-bfe7a6754c05" containerName="keystone-bootstrap" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671308 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671319 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671328 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdcab93e-7564-4d7e-b4fb-2da47c638c4e" containerName="init" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.671988 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.674198 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.674952 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.675829 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9jg9" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.676006 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.676144 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.682202 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b5bbr"] Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.796935 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.797002 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.797128 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.797235 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfh5f\" (UniqueName: \"kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.797348 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.797512 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899645 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899742 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899767 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899808 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfh5f\" (UniqueName: \"kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899854 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.899883 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.905672 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.905885 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.909922 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.916515 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.917153 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.922567 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfh5f\" (UniqueName: \"kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f\") pod \"keystone-bootstrap-b5bbr\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:11 crc kubenswrapper[4867]: I1212 07:08:11.992544 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.132914 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.132914 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" event={"ID":"94510a33-be0c-4f99-8fee-ad22a1b86719","Type":"ContainerDied","Data":"7f3f05b8058bb51f6686820e54789ed0b54cbe51b9583abd0e4a16b408869122"} Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.169373 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.178028 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8db84466c-9tz5f"] Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.298903 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8db84466c-9tz5f" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 12 07:08:12 crc kubenswrapper[4867]: E1212 07:08:12.825949 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:33f4e5f7a715d48482ec46a42267ea992fa268585303c4f1bd3cbea072a6348b" Dec 12 07:08:12 crc kubenswrapper[4867]: E1212 07:08:12.826691 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:33f4e5f7a715d48482ec46a42267ea992fa268585303c4f1bd3cbea072a6348b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pck7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-6lmsp_openstack(7299cd15-ce08-48f3-be09-c3a70901c4aa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 07:08:12 crc kubenswrapper[4867]: E1212 07:08:12.828174 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-6lmsp" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.848927 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94510a33-be0c-4f99-8fee-ad22a1b86719" path="/var/lib/kubelet/pods/94510a33-be0c-4f99-8fee-ad22a1b86719/volumes" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.849690 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" path="/var/lib/kubelet/pods/b0f75dd4-02a1-473c-a06f-c8477baa2d90/volumes" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.850254 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d501d207-060c-4e80-b631-bfe7a6754c05" path="/var/lib/kubelet/pods/d501d207-060c-4e80-b631-bfe7a6754c05/volumes" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.874793 4867 scope.go:117] "RemoveContainer" containerID="63bf09d8ac7be5ac381adb2e7a7413967e43e8b8f5634ed64f16a2bfcf37a4d4" Dec 12 07:08:12 crc kubenswrapper[4867]: I1212 07:08:12.911905 4867 scope.go:117] "RemoveContainer" containerID="9093b22e58464b9a68507926c5e4eef2f36c414ba9a29175fac943e3c3501e62" Dec 12 07:08:13 crc kubenswrapper[4867]: I1212 07:08:13.145182 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9dbsw" event={"ID":"68801d28-d85f-46b6-86cd-debaf6dab7ca","Type":"ContainerStarted","Data":"38a25c4bf9c0c0aa9898ef57ec5198b23b894a9894b5b11dd705553353f23d8d"} Dec 12 07:08:13 crc kubenswrapper[4867]: E1212 07:08:13.148864 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:33f4e5f7a715d48482ec46a42267ea992fa268585303c4f1bd3cbea072a6348b\\\"\"" pod="openstack/placement-db-sync-6lmsp" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" Dec 12 07:08:13 crc kubenswrapper[4867]: I1212 07:08:13.165532 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9dbsw" podStartSLOduration=1.485013302 podStartE2EDuration="38.165481497s" podCreationTimestamp="2025-12-12 07:07:35 +0000 UTC" firstStartedPulling="2025-12-12 07:07:36.207403268 +0000 UTC m=+1143.778784527" lastFinishedPulling="2025-12-12 07:08:12.887871453 +0000 UTC m=+1180.459252722" observedRunningTime="2025-12-12 07:08:13.164524685 +0000 UTC m=+1180.735905954" watchObservedRunningTime="2025-12-12 07:08:13.165481497 +0000 UTC m=+1180.736862756" Dec 12 07:08:13 crc kubenswrapper[4867]: I1212 07:08:13.306723 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b5bbr"] Dec 12 07:08:13 crc kubenswrapper[4867]: W1212 07:08:13.311501 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b60ab40_1d50_4c2d_a9f3_726f0825c58a.slice/crio-cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246 WatchSource:0}: Error finding container cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246: Status 404 returned error can't find the container with id cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246 Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.158152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b5bbr" event={"ID":"3b60ab40-1d50-4c2d-a9f3-726f0825c58a","Type":"ContainerStarted","Data":"fb810476f4594e24532a01d472f2d4a6803aa3cc6fe3bdf084703be4a759b4ce"} Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.158807 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b5bbr" event={"ID":"3b60ab40-1d50-4c2d-a9f3-726f0825c58a","Type":"ContainerStarted","Data":"cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246"} Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.159832 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerStarted","Data":"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3"} Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.161738 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3a5237b-ffff-43b1-a921-68affabec6cd" containerID="bcf8eb7e4e5ab4fec8f5c52187980849509ed93d855f00433e7a743ea90ffb05" exitCode=0 Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.161776 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7lxjs" event={"ID":"a3a5237b-ffff-43b1-a921-68affabec6cd","Type":"ContainerDied","Data":"bcf8eb7e4e5ab4fec8f5c52187980849509ed93d855f00433e7a743ea90ffb05"} Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.180409 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b5bbr" podStartSLOduration=3.180390645 podStartE2EDuration="3.180390645s" podCreationTimestamp="2025-12-12 07:08:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:14.172833859 +0000 UTC m=+1181.744215138" watchObservedRunningTime="2025-12-12 07:08:14.180390645 +0000 UTC m=+1181.751771914" Dec 12 07:08:14 crc kubenswrapper[4867]: I1212 07:08:14.355826 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67fdf7998c-rlltl" podUID="b0f75dd4-02a1-473c-a06f-c8477baa2d90" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: i/o timeout" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.195295 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7lxjs" event={"ID":"a3a5237b-ffff-43b1-a921-68affabec6cd","Type":"ContainerDied","Data":"62708c7aca8968aaf36f542b839b98f234c61924be7fa50ef39ff756fb422dd3"} Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.195762 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62708c7aca8968aaf36f542b839b98f234c61924be7fa50ef39ff756fb422dd3" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.221161 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7lxjs" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.310404 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle\") pod \"a3a5237b-ffff-43b1-a921-68affabec6cd\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.310532 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data\") pod \"a3a5237b-ffff-43b1-a921-68affabec6cd\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.310640 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbndn\" (UniqueName: \"kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn\") pod \"a3a5237b-ffff-43b1-a921-68affabec6cd\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.310717 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data\") pod \"a3a5237b-ffff-43b1-a921-68affabec6cd\" (UID: \"a3a5237b-ffff-43b1-a921-68affabec6cd\") " Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.318562 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a3a5237b-ffff-43b1-a921-68affabec6cd" (UID: "a3a5237b-ffff-43b1-a921-68affabec6cd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.319517 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn" (OuterVolumeSpecName: "kube-api-access-bbndn") pod "a3a5237b-ffff-43b1-a921-68affabec6cd" (UID: "a3a5237b-ffff-43b1-a921-68affabec6cd"). InnerVolumeSpecName "kube-api-access-bbndn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.343109 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3a5237b-ffff-43b1-a921-68affabec6cd" (UID: "a3a5237b-ffff-43b1-a921-68affabec6cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.374188 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data" (OuterVolumeSpecName: "config-data") pod "a3a5237b-ffff-43b1-a921-68affabec6cd" (UID: "a3a5237b-ffff-43b1-a921-68affabec6cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.412812 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.412855 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.412866 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbndn\" (UniqueName: \"kubernetes.io/projected/a3a5237b-ffff-43b1-a921-68affabec6cd-kube-api-access-bbndn\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:17 crc kubenswrapper[4867]: I1212 07:08:17.412879 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3a5237b-ffff-43b1-a921-68affabec6cd-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.206306 4867 generic.go:334] "Generic (PLEG): container finished" podID="68801d28-d85f-46b6-86cd-debaf6dab7ca" containerID="38a25c4bf9c0c0aa9898ef57ec5198b23b894a9894b5b11dd705553353f23d8d" exitCode=0 Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.206491 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9dbsw" event={"ID":"68801d28-d85f-46b6-86cd-debaf6dab7ca","Type":"ContainerDied","Data":"38a25c4bf9c0c0aa9898ef57ec5198b23b894a9894b5b11dd705553353f23d8d"} Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.208801 4867 generic.go:334] "Generic (PLEG): container finished" podID="3b60ab40-1d50-4c2d-a9f3-726f0825c58a" containerID="fb810476f4594e24532a01d472f2d4a6803aa3cc6fe3bdf084703be4a759b4ce" exitCode=0 Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.208885 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7lxjs" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.209269 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b5bbr" event={"ID":"3b60ab40-1d50-4c2d-a9f3-726f0825c58a","Type":"ContainerDied","Data":"fb810476f4594e24532a01d472f2d4a6803aa3cc6fe3bdf084703be4a759b4ce"} Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.642078 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:18 crc kubenswrapper[4867]: E1212 07:08:18.642532 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" containerName="glance-db-sync" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.642552 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" containerName="glance-db-sync" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.642803 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" containerName="glance-db-sync" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.643915 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.662260 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746101 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746175 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746195 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746267 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hbvk\" (UniqueName: \"kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746297 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.746340 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847276 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847354 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847377 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847435 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hbvk\" (UniqueName: \"kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847466 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.847512 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.848447 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.849309 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.849594 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.849873 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.850835 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.868806 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hbvk\" (UniqueName: \"kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk\") pod \"dnsmasq-dns-6f6f8cb849-bhgf2\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:18 crc kubenswrapper[4867]: I1212 07:08:18.967637 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.550637 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.552732 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.554982 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lggbv" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.555249 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.556798 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.574458 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671470 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671574 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671594 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671635 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671680 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671707 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.671745 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gx4q\" (UniqueName: \"kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773141 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773216 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773257 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gx4q\" (UniqueName: \"kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773315 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773395 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773420 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.773592 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.774077 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.774124 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.774279 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.780903 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.781858 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.783427 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.796741 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gx4q\" (UniqueName: \"kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.798727 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.800306 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.806832 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.834003 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.862758 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877076 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877286 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877392 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltd7c\" (UniqueName: \"kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877571 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877648 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877735 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.877833 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.898207 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980302 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980431 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltd7c\" (UniqueName: \"kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980555 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980629 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.980736 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.981881 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.982100 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.982502 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.987203 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.988456 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.989281 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:19 crc kubenswrapper[4867]: I1212 07:08:19.996604 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltd7c\" (UniqueName: \"kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:20 crc kubenswrapper[4867]: I1212 07:08:20.025863 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:20 crc kubenswrapper[4867]: I1212 07:08:20.210735 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:21 crc kubenswrapper[4867]: I1212 07:08:21.029902 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:21 crc kubenswrapper[4867]: I1212 07:08:21.089796 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.733207 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.788077 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data\") pod \"68801d28-d85f-46b6-86cd-debaf6dab7ca\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.788485 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqc6f\" (UniqueName: \"kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f\") pod \"68801d28-d85f-46b6-86cd-debaf6dab7ca\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.788535 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle\") pod \"68801d28-d85f-46b6-86cd-debaf6dab7ca\" (UID: \"68801d28-d85f-46b6-86cd-debaf6dab7ca\") " Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.794721 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "68801d28-d85f-46b6-86cd-debaf6dab7ca" (UID: "68801d28-d85f-46b6-86cd-debaf6dab7ca"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.794812 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f" (OuterVolumeSpecName: "kube-api-access-fqc6f") pod "68801d28-d85f-46b6-86cd-debaf6dab7ca" (UID: "68801d28-d85f-46b6-86cd-debaf6dab7ca"). InnerVolumeSpecName "kube-api-access-fqc6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.814641 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68801d28-d85f-46b6-86cd-debaf6dab7ca" (UID: "68801d28-d85f-46b6-86cd-debaf6dab7ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.891007 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.891034 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqc6f\" (UniqueName: \"kubernetes.io/projected/68801d28-d85f-46b6-86cd-debaf6dab7ca-kube-api-access-fqc6f\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:25 crc kubenswrapper[4867]: I1212 07:08:25.891048 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68801d28-d85f-46b6-86cd-debaf6dab7ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:26 crc kubenswrapper[4867]: I1212 07:08:26.277631 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9dbsw" event={"ID":"68801d28-d85f-46b6-86cd-debaf6dab7ca","Type":"ContainerDied","Data":"19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784"} Dec 12 07:08:26 crc kubenswrapper[4867]: I1212 07:08:26.277675 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19b140fbcd08d2b1438b48123c537dfdfa257c74eaf244fc0ab3ebc1fb4e8784" Dec 12 07:08:26 crc kubenswrapper[4867]: I1212 07:08:26.277675 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9dbsw" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:26.999896 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:08:27 crc kubenswrapper[4867]: E1212 07:08:27.000476 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" containerName="barbican-db-sync" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.000494 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" containerName="barbican-db-sync" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.000722 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" containerName="barbican-db-sync" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.001722 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.006650 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.006993 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.007528 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-qqxfz" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.024115 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.027789 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.031829 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.046001 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.072422 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115425 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115526 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115560 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115640 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115674 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115702 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115734 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115766 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5tfh\" (UniqueName: \"kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115802 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w29jc\" (UniqueName: \"kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.115899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.126527 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.163257 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.164785 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.199886 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.217311 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.217396 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.224400 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.224803 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.224853 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.224894 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225178 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225300 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225345 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5tfh\" (UniqueName: \"kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225376 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w29jc\" (UniqueName: \"kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225473 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225490 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225592 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225668 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225728 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26h6f\" (UniqueName: \"kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225787 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.225808 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.227157 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.230858 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.233307 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.233955 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.235664 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.242446 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.245065 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.249038 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w29jc\" (UniqueName: \"kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc\") pod \"barbican-keystone-listener-b4475fc66-dqrh7\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.274914 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5tfh\" (UniqueName: \"kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh\") pod \"barbican-worker-6d5d4d56b5-6h6kt\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.299782 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.302748 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.309014 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.322887 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.324359 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327465 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327509 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327535 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327567 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327596 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327614 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327633 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327682 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2sd8\" (UniqueName: \"kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327714 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327749 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.327776 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26h6f\" (UniqueName: \"kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.335446 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.364522 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.429291 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.429367 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.429403 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.429453 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2sd8\" (UniqueName: \"kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.429484 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.790367 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.790730 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.790760 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.791048 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.791075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.794673 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.795062 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.795320 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.795485 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2sd8\" (UniqueName: \"kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8\") pod \"barbican-api-579b864c64-lrpn7\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:27 crc kubenswrapper[4867]: I1212 07:08:27.796918 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26h6f\" (UniqueName: \"kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f\") pod \"dnsmasq-dns-8fffc8985-2v9wn\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.090622 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.092828 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.726530 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753330 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753523 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753565 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfh5f\" (UniqueName: \"kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753618 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753740 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.753807 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.764269 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.764317 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f" (OuterVolumeSpecName: "kube-api-access-lfh5f") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "kube-api-access-lfh5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.764576 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.777468 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts" (OuterVolumeSpecName: "scripts") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:28 crc kubenswrapper[4867]: E1212 07:08:28.780675 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle podName:3b60ab40-1d50-4c2d-a9f3-726f0825c58a nodeName:}" failed. No retries permitted until 2025-12-12 07:08:29.280604972 +0000 UTC m=+1196.851986311 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a") : error deleting /var/lib/kubelet/pods/3b60ab40-1d50-4c2d-a9f3-726f0825c58a/volume-subpaths: remove /var/lib/kubelet/pods/3b60ab40-1d50-4c2d-a9f3-726f0825c58a/volume-subpaths: no such file or directory Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.803949 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data" (OuterVolumeSpecName: "config-data") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.856535 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.856561 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.856571 4867 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.856579 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.856587 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfh5f\" (UniqueName: \"kubernetes.io/projected/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-kube-api-access-lfh5f\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.988931 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:08:28 crc kubenswrapper[4867]: I1212 07:08:28.988984 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.188364 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.314669 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b5bbr" event={"ID":"3b60ab40-1d50-4c2d-a9f3-726f0825c58a","Type":"ContainerDied","Data":"cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246"} Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.315030 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cef9495970ad3823630b3bb10f82fd3c4976131d153d8cca9a7becebc87ca246" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.314674 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b5bbr" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.315083 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.367418 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") pod \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\" (UID: \"3b60ab40-1d50-4c2d-a9f3-726f0825c58a\") " Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.377642 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b60ab40-1d50-4c2d-a9f3-726f0825c58a" (UID: "3b60ab40-1d50-4c2d-a9f3-726f0825c58a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.477520 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b60ab40-1d50-4c2d-a9f3-726f0825c58a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.683107 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.830617 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.839296 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.915689 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:08:29 crc kubenswrapper[4867]: E1212 07:08:29.916134 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b60ab40-1d50-4c2d-a9f3-726f0825c58a" containerName="keystone-bootstrap" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.916151 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b60ab40-1d50-4c2d-a9f3-726f0825c58a" containerName="keystone-bootstrap" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.916420 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b60ab40-1d50-4c2d-a9f3-726f0825c58a" containerName="keystone-bootstrap" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.917203 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.933888 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.934170 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.934186 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-s9jg9" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.934943 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.935151 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.935392 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.943996 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988621 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988658 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988705 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rh9\" (UniqueName: \"kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988801 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988824 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:29 crc kubenswrapper[4867]: I1212 07:08:29.988869 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: W1212 07:08:30.016261 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06e70b8d_199d_4c5b_be9c_7edc6daf4bc8.slice/crio-5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee WatchSource:0}: Error finding container 5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee: Status 404 returned error can't find the container with id 5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.020098 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.097271 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rh9\" (UniqueName: \"kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.098103 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.098193 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.098278 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.098385 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.102673 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.108848 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.109017 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.108771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.119937 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.120099 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.120624 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.127044 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.128002 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.128665 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rh9\" (UniqueName: \"kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.130195 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs\") pod \"keystone-74796bf6db-754rv\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.145712 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.168556 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.180583 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.183719 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.183909 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.194209 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326489 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326575 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326622 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326678 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326744 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5fmv\" (UniqueName: \"kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.326787 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.331172 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429340 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429412 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429500 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429554 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429625 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5fmv\" (UniqueName: \"kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.429669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.431643 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.438531 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.445860 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.448929 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.449165 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.457778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.461664 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" event={"ID":"03a0f30d-fe35-4e6b-bdfd-746f1040c82b","Type":"ContainerStarted","Data":"8f579aef811a9e4d9a2e7f3f30be0188f7311eb730690bced8c1cb5f44eb33e9"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.475977 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5fmv\" (UniqueName: \"kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv\") pod \"barbican-api-6f69644fbd-dsggh\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.503016 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerStarted","Data":"5cc2f57e55a876dcce8294906fe1abe660571988a9a7e8e34b76a4656e3337ef"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.510548 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerStarted","Data":"7326846abd30f5952886dc11fac8920a0f4cc06d67c373a38068cd5913c8d8dd"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.516850 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerStarted","Data":"9e7b76644371ac1d38436f658179e76b6a5d2acda57f631db86c338a98275947"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.526359 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lmsp" event={"ID":"7299cd15-ce08-48f3-be09-c3a70901c4aa","Type":"ContainerStarted","Data":"ba666498132b53c649f4905c4a572219fb0ab53120aad5d06225092d208e3c0d"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.529695 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerStarted","Data":"5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.551790 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.556441 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6lmsp" podStartSLOduration=2.801383959 podStartE2EDuration="55.556420862s" podCreationTimestamp="2025-12-12 07:07:35 +0000 UTC" firstStartedPulling="2025-12-12 07:07:36.602111783 +0000 UTC m=+1144.173493042" lastFinishedPulling="2025-12-12 07:08:29.357148676 +0000 UTC m=+1196.928529945" observedRunningTime="2025-12-12 07:08:30.548016875 +0000 UTC m=+1198.119398144" watchObservedRunningTime="2025-12-12 07:08:30.556420862 +0000 UTC m=+1198.127802131" Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.580489 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerStarted","Data":"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.589286 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerStarted","Data":"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.589327 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerStarted","Data":"e698be70a47300a227af9224e30aa7c195bada2a74cb763728088013433ec5b9"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.593712 4867 generic.go:334] "Generic (PLEG): container finished" podID="61c7bead-9414-4649-a277-97ea959a489d" containerID="11161a424ea7f257e000952382f1929871aed5a7f741c19a335f810ef69cb41e" exitCode=0 Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.593749 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" event={"ID":"61c7bead-9414-4649-a277-97ea959a489d","Type":"ContainerDied","Data":"11161a424ea7f257e000952382f1929871aed5a7f741c19a335f810ef69cb41e"} Dec 12 07:08:30 crc kubenswrapper[4867]: I1212 07:08:30.593771 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" event={"ID":"61c7bead-9414-4649-a277-97ea959a489d","Type":"ContainerStarted","Data":"f0e7c6fffbd281391f139f2b83b0e3b41f38e460767530cd38ace96c4f647723"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.114322 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:08:31 crc kubenswrapper[4867]: W1212 07:08:31.123503 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38431807_5fbd_4f45_8eca_b42af2c0d221.slice/crio-6165b97c8c408c0ad06de5b242132547e47ddecd324d40e2800e468f4b8d4f8c WatchSource:0}: Error finding container 6165b97c8c408c0ad06de5b242132547e47ddecd324d40e2800e468f4b8d4f8c: Status 404 returned error can't find the container with id 6165b97c8c408c0ad06de5b242132547e47ddecd324d40e2800e468f4b8d4f8c Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.136388 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248048 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248177 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248207 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248374 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hbvk\" (UniqueName: \"kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248428 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.248596 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0\") pod \"61c7bead-9414-4649-a277-97ea959a489d\" (UID: \"61c7bead-9414-4649-a277-97ea959a489d\") " Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.257412 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk" (OuterVolumeSpecName: "kube-api-access-9hbvk") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "kube-api-access-9hbvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.284069 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.293913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.294000 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.295488 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.295667 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config" (OuterVolumeSpecName: "config") pod "61c7bead-9414-4649-a277-97ea959a489d" (UID: "61c7bead-9414-4649-a277-97ea959a489d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358708 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358756 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358768 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358780 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358792 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hbvk\" (UniqueName: \"kubernetes.io/projected/61c7bead-9414-4649-a277-97ea959a489d-kube-api-access-9hbvk\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.358807 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61c7bead-9414-4649-a277-97ea959a489d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.378772 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.611162 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerStarted","Data":"f0f285c377bf349906dceb0496200794f517f7950aa4395ff3b67bad15a203f6"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.614918 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74796bf6db-754rv" event={"ID":"38431807-5fbd-4f45-8eca-b42af2c0d221","Type":"ContainerStarted","Data":"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.614963 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74796bf6db-754rv" event={"ID":"38431807-5fbd-4f45-8eca-b42af2c0d221","Type":"ContainerStarted","Data":"6165b97c8c408c0ad06de5b242132547e47ddecd324d40e2800e468f4b8d4f8c"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.615848 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.618583 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7fnd8" event={"ID":"63341abf-5801-4239-8370-4c0732c95e3b","Type":"ContainerStarted","Data":"3f67093be35c467d200dbddfac2e3e8c6dc0d42da9b5d84d7d84b0c2b0ca891d"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.621421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerStarted","Data":"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.621509 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.621541 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.623677 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.623676 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6f8cb849-bhgf2" event={"ID":"61c7bead-9414-4649-a277-97ea959a489d","Type":"ContainerDied","Data":"f0e7c6fffbd281391f139f2b83b0e3b41f38e460767530cd38ace96c4f647723"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.623838 4867 scope.go:117] "RemoveContainer" containerID="11161a424ea7f257e000952382f1929871aed5a7f741c19a335f810ef69cb41e" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.627269 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerStarted","Data":"cb590d6285be23ac892b56c5aef2863e103571d677ab794adca39f7fb7e606fc"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.628228 4867 generic.go:334] "Generic (PLEG): container finished" podID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerID="7608d1d79d129f76784e50627ed23dd3a9f295ec2ef5e27466b66734c863a038" exitCode=0 Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.628334 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" event={"ID":"03a0f30d-fe35-4e6b-bdfd-746f1040c82b","Type":"ContainerDied","Data":"7608d1d79d129f76784e50627ed23dd3a9f295ec2ef5e27466b66734c863a038"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.631813 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerStarted","Data":"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051"} Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.667654 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-74796bf6db-754rv" podStartSLOduration=2.6676356009999997 podStartE2EDuration="2.667635601s" podCreationTimestamp="2025-12-12 07:08:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:31.64893841 +0000 UTC m=+1199.220319709" watchObservedRunningTime="2025-12-12 07:08:31.667635601 +0000 UTC m=+1199.239016870" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.684672 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-579b864c64-lrpn7" podStartSLOduration=4.681216884 podStartE2EDuration="4.681216884s" podCreationTimestamp="2025-12-12 07:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:31.681084351 +0000 UTC m=+1199.252465640" watchObservedRunningTime="2025-12-12 07:08:31.681216884 +0000 UTC m=+1199.252598143" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.800826 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7fnd8" podStartSLOduration=3.76579266 podStartE2EDuration="56.800808079s" podCreationTimestamp="2025-12-12 07:07:35 +0000 UTC" firstStartedPulling="2025-12-12 07:07:36.324788422 +0000 UTC m=+1143.896169681" lastFinishedPulling="2025-12-12 07:08:29.359803831 +0000 UTC m=+1196.931185100" observedRunningTime="2025-12-12 07:08:31.745397435 +0000 UTC m=+1199.316778704" watchObservedRunningTime="2025-12-12 07:08:31.800808079 +0000 UTC m=+1199.372189348" Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.831819 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:31 crc kubenswrapper[4867]: I1212 07:08:31.856054 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f6f8cb849-bhgf2"] Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.645453 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerStarted","Data":"3de038c21eac21d9ff13f273987dcdd7a3a457a5962ccda343b200278c02f4ac"} Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.645541 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-log" containerID="cri-o://f0f285c377bf349906dceb0496200794f517f7950aa4395ff3b67bad15a203f6" gracePeriod=30 Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.645650 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-httpd" containerID="cri-o://3de038c21eac21d9ff13f273987dcdd7a3a457a5962ccda343b200278c02f4ac" gracePeriod=30 Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.653822 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerStarted","Data":"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b"} Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.676131 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=14.676110479 podStartE2EDuration="14.676110479s" podCreationTimestamp="2025-12-12 07:08:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:32.669811893 +0000 UTC m=+1200.241193172" watchObservedRunningTime="2025-12-12 07:08:32.676110479 +0000 UTC m=+1200.247491748" Dec 12 07:08:32 crc kubenswrapper[4867]: I1212 07:08:32.855352 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c7bead-9414-4649-a277-97ea959a489d" path="/var/lib/kubelet/pods/61c7bead-9414-4649-a277-97ea959a489d/volumes" Dec 12 07:08:33 crc kubenswrapper[4867]: I1212 07:08:33.665243 4867 generic.go:334] "Generic (PLEG): container finished" podID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerID="3de038c21eac21d9ff13f273987dcdd7a3a457a5962ccda343b200278c02f4ac" exitCode=0 Dec 12 07:08:33 crc kubenswrapper[4867]: I1212 07:08:33.665658 4867 generic.go:334] "Generic (PLEG): container finished" podID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerID="f0f285c377bf349906dceb0496200794f517f7950aa4395ff3b67bad15a203f6" exitCode=143 Dec 12 07:08:33 crc kubenswrapper[4867]: I1212 07:08:33.665351 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerDied","Data":"3de038c21eac21d9ff13f273987dcdd7a3a457a5962ccda343b200278c02f4ac"} Dec 12 07:08:33 crc kubenswrapper[4867]: I1212 07:08:33.665701 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerDied","Data":"f0f285c377bf349906dceb0496200794f517f7950aa4395ff3b67bad15a203f6"} Dec 12 07:08:34 crc kubenswrapper[4867]: I1212 07:08:34.675692 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-log" containerID="cri-o://917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" gracePeriod=30 Dec 12 07:08:34 crc kubenswrapper[4867]: I1212 07:08:34.676302 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-httpd" containerID="cri-o://ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" gracePeriod=30 Dec 12 07:08:34 crc kubenswrapper[4867]: I1212 07:08:34.711191 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=16.711170641 podStartE2EDuration="16.711170641s" podCreationTimestamp="2025-12-12 07:08:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:34.698182982 +0000 UTC m=+1202.269564251" watchObservedRunningTime="2025-12-12 07:08:34.711170641 +0000 UTC m=+1202.282551910" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.209296 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.305991 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306044 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306090 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306176 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306341 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gx4q\" (UniqueName: \"kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306496 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data\") pod \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\" (UID: \"c91e7ebb-06cb-4210-ae1a-870b20017fb5\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.306779 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs" (OuterVolumeSpecName: "logs") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.307484 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.307563 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.318749 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts" (OuterVolumeSpecName: "scripts") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.318998 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q" (OuterVolumeSpecName: "kube-api-access-6gx4q") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "kube-api-access-6gx4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.344895 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.411442 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c91e7ebb-06cb-4210-ae1a-870b20017fb5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.411488 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.411522 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.411537 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gx4q\" (UniqueName: \"kubernetes.io/projected/c91e7ebb-06cb-4210-ae1a-870b20017fb5-kube-api-access-6gx4q\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.439752 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.458135 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.469312 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data" (OuterVolumeSpecName: "config-data") pod "c91e7ebb-06cb-4210-ae1a-870b20017fb5" (UID: "c91e7ebb-06cb-4210-ae1a-870b20017fb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.513784 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.513830 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.513843 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c91e7ebb-06cb-4210-ae1a-870b20017fb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.672486 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.690240 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c91e7ebb-06cb-4210-ae1a-870b20017fb5","Type":"ContainerDied","Data":"7326846abd30f5952886dc11fac8920a0f4cc06d67c373a38068cd5913c8d8dd"} Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.690272 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.690298 4867 scope.go:117] "RemoveContainer" containerID="3de038c21eac21d9ff13f273987dcdd7a3a457a5962ccda343b200278c02f4ac" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.702256 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerStarted","Data":"e77048aeb4e616861c5d278c417e28346695e6edbd2cfceba516f723cc72bdf9"} Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705423 4867 generic.go:334] "Generic (PLEG): container finished" podID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerID="ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" exitCode=0 Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705465 4867 generic.go:334] "Generic (PLEG): container finished" podID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerID="917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" exitCode=143 Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705528 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerDied","Data":"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b"} Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705559 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerDied","Data":"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051"} Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705572 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a5c1ac74-4f32-4540-a14b-e5c37f82fc29","Type":"ContainerDied","Data":"5cc2f57e55a876dcce8294906fe1abe660571988a9a7e8e34b76a4656e3337ef"} Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.705898 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.724900 4867 scope.go:117] "RemoveContainer" containerID="f0f285c377bf349906dceb0496200794f517f7950aa4395ff3b67bad15a203f6" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.744281 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.777571 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.782642 4867 scope.go:117] "RemoveContainer" containerID="ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.808601 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:35 crc kubenswrapper[4867]: E1212 07:08:35.809069 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809082 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: E1212 07:08:35.809097 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809103 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: E1212 07:08:35.809113 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809120 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: E1212 07:08:35.809128 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809134 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: E1212 07:08:35.809146 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c7bead-9414-4649-a277-97ea959a489d" containerName="init" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809151 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c7bead-9414-4649-a277-97ea959a489d" containerName="init" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809334 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809356 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c7bead-9414-4649-a277-97ea959a489d" containerName="init" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809365 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809373 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-httpd" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.809385 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" containerName="glance-log" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.810322 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.813508 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.815871 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819158 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819411 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819467 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819517 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819626 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltd7c\" (UniqueName: \"kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819699 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.819788 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\" (UID: \"a5c1ac74-4f32-4540-a14b-e5c37f82fc29\") " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.820449 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs" (OuterVolumeSpecName: "logs") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.820704 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.821885 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.823935 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts" (OuterVolumeSpecName: "scripts") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.825159 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c" (OuterVolumeSpecName: "kube-api-access-ltd7c") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "kube-api-access-ltd7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.826600 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.832259 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.852601 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.896517 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data" (OuterVolumeSpecName: "config-data") pod "a5c1ac74-4f32-4540-a14b-e5c37f82fc29" (UID: "a5c1ac74-4f32-4540-a14b-e5c37f82fc29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923078 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923261 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923309 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923345 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923370 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gphg6\" (UniqueName: \"kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923408 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923751 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923836 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923850 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923859 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltd7c\" (UniqueName: \"kubernetes.io/projected/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-kube-api-access-ltd7c\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923868 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923887 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.923898 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5c1ac74-4f32-4540-a14b-e5c37f82fc29-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.950993 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 12 07:08:35 crc kubenswrapper[4867]: I1212 07:08:35.962630 4867 scope.go:117] "RemoveContainer" containerID="917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.000088 4867 scope.go:117] "RemoveContainer" containerID="ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" Dec 12 07:08:36 crc kubenswrapper[4867]: E1212 07:08:36.000806 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b\": container with ID starting with ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b not found: ID does not exist" containerID="ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.000873 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b"} err="failed to get container status \"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b\": rpc error: code = NotFound desc = could not find container \"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b\": container with ID starting with ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b not found: ID does not exist" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.000908 4867 scope.go:117] "RemoveContainer" containerID="917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" Dec 12 07:08:36 crc kubenswrapper[4867]: E1212 07:08:36.002315 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051\": container with ID starting with 917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051 not found: ID does not exist" containerID="917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.002469 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051"} err="failed to get container status \"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051\": rpc error: code = NotFound desc = could not find container \"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051\": container with ID starting with 917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051 not found: ID does not exist" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.002519 4867 scope.go:117] "RemoveContainer" containerID="ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.003085 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b"} err="failed to get container status \"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b\": rpc error: code = NotFound desc = could not find container \"ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b\": container with ID starting with ef8b1a109a75de437af60d0f548d378757b6ebd78c8a10497e818933b9a89b0b not found: ID does not exist" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.003119 4867 scope.go:117] "RemoveContainer" containerID="917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.003865 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051"} err="failed to get container status \"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051\": rpc error: code = NotFound desc = could not find container \"917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051\": container with ID starting with 917f0d0d41bb41143e9dbcd6b02887a60b14349a47143327a4f83e4f0c531051 not found: ID does not exist" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025595 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025648 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gphg6\" (UniqueName: \"kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025674 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025694 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025747 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025860 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025907 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.025987 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.026454 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.028073 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.028130 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.033213 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.041924 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.044039 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.045952 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.050647 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gphg6\" (UniqueName: \"kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.051375 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.067138 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.094297 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.096395 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.106276 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.106630 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.110314 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.130969 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.229757 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtdx8\" (UniqueName: \"kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.229809 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.229850 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.229884 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.230138 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.230244 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.230384 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.230579 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.302583 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.332381 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtdx8\" (UniqueName: \"kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.332438 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333125 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333189 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333315 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333383 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333428 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.333834 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.334530 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.334887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.344982 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.345949 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.345985 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.347379 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.356098 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtdx8\" (UniqueName: \"kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.414752 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.430143 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.724731 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerStarted","Data":"0482c34d8c8358e81c2f8bc9ca4911f7c07f4fe5e94574f9a0a5812a8529427f"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.725148 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerStarted","Data":"1499f0119bb5be4db9007f30932c9a39e2288f464937e17ccb75934aea9e9680"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.728000 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerStarted","Data":"17557e0767d1d09b0c76be9d8ba0d9223b9b9565bc8ce63ba2fb3d24cdeacbe1"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.729026 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.729059 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.751737 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" event={"ID":"03a0f30d-fe35-4e6b-bdfd-746f1040c82b","Type":"ContainerStarted","Data":"86e3b995214e4f591f726252c55ce30449f93230f261366531b28c1a122cd8c5"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.752742 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.758617 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerStarted","Data":"23284a0c2e794f7efbc84dbb67df86a5355b8c4e5689bfc88bd6bad7d0096872"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.758659 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerStarted","Data":"14a5c4be4116f44371426de020f2bddb11420dd71b7ecd0f938dff04595d0193"} Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.760513 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" podStartSLOduration=5.316804013 podStartE2EDuration="10.760492226s" podCreationTimestamp="2025-12-12 07:08:26 +0000 UTC" firstStartedPulling="2025-12-12 07:08:29.860892918 +0000 UTC m=+1197.432274187" lastFinishedPulling="2025-12-12 07:08:35.304581131 +0000 UTC m=+1202.875962400" observedRunningTime="2025-12-12 07:08:36.745789015 +0000 UTC m=+1204.317170294" watchObservedRunningTime="2025-12-12 07:08:36.760492226 +0000 UTC m=+1204.331873495" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.773514 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6f69644fbd-dsggh" podStartSLOduration=6.773497376 podStartE2EDuration="6.773497376s" podCreationTimestamp="2025-12-12 07:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:36.766548505 +0000 UTC m=+1204.337929774" watchObservedRunningTime="2025-12-12 07:08:36.773497376 +0000 UTC m=+1204.344878645" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.815533 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" podStartSLOduration=9.81550766 podStartE2EDuration="9.81550766s" podCreationTimestamp="2025-12-12 07:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:36.8004332 +0000 UTC m=+1204.371814569" watchObservedRunningTime="2025-12-12 07:08:36.81550766 +0000 UTC m=+1204.386888929" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.853686 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c1ac74-4f32-4540-a14b-e5c37f82fc29" path="/var/lib/kubelet/pods/a5c1ac74-4f32-4540-a14b-e5c37f82fc29/volumes" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.856866 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c91e7ebb-06cb-4210-ae1a-870b20017fb5" path="/var/lib/kubelet/pods/c91e7ebb-06cb-4210-ae1a-870b20017fb5/volumes" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.940622 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" podStartSLOduration=5.661630322 podStartE2EDuration="10.940602371s" podCreationTimestamp="2025-12-12 07:08:26 +0000 UTC" firstStartedPulling="2025-12-12 07:08:30.024917916 +0000 UTC m=+1197.596299185" lastFinishedPulling="2025-12-12 07:08:35.303889965 +0000 UTC m=+1202.875271234" observedRunningTime="2025-12-12 07:08:36.820954175 +0000 UTC m=+1204.392335464" watchObservedRunningTime="2025-12-12 07:08:36.940602371 +0000 UTC m=+1204.511983640" Dec 12 07:08:36 crc kubenswrapper[4867]: I1212 07:08:36.951282 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:08:36 crc kubenswrapper[4867]: W1212 07:08:36.980703 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3844a91f_5305_40f8_80a8_7801706fe0b4.slice/crio-52314552ac36652bea8ffdce9e8c5ddb8e12c6d6a7bd8adb8811dc4750fc41d2 WatchSource:0}: Error finding container 52314552ac36652bea8ffdce9e8c5ddb8e12c6d6a7bd8adb8811dc4750fc41d2: Status 404 returned error can't find the container with id 52314552ac36652bea8ffdce9e8c5ddb8e12c6d6a7bd8adb8811dc4750fc41d2 Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.048206 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:08:37 crc kubenswrapper[4867]: W1212 07:08:37.064596 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf77f99b8_46c3_4b6d_abe7_f6ce348f4d72.slice/crio-adf18422e18f264a1ef24f3681bcf076a5e9b573c0330fe953766a004768d5be WatchSource:0}: Error finding container adf18422e18f264a1ef24f3681bcf076a5e9b573c0330fe953766a004768d5be: Status 404 returned error can't find the container with id adf18422e18f264a1ef24f3681bcf076a5e9b573c0330fe953766a004768d5be Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.783659 4867 generic.go:334] "Generic (PLEG): container finished" podID="7299cd15-ce08-48f3-be09-c3a70901c4aa" containerID="ba666498132b53c649f4905c4a572219fb0ab53120aad5d06225092d208e3c0d" exitCode=0 Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.783947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lmsp" event={"ID":"7299cd15-ce08-48f3-be09-c3a70901c4aa","Type":"ContainerDied","Data":"ba666498132b53c649f4905c4a572219fb0ab53120aad5d06225092d208e3c0d"} Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.785944 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerStarted","Data":"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6"} Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.785971 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerStarted","Data":"52314552ac36652bea8ffdce9e8c5ddb8e12c6d6a7bd8adb8811dc4750fc41d2"} Dec 12 07:08:37 crc kubenswrapper[4867]: I1212 07:08:37.788714 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerStarted","Data":"adf18422e18f264a1ef24f3681bcf076a5e9b573c0330fe953766a004768d5be"} Dec 12 07:08:38 crc kubenswrapper[4867]: I1212 07:08:38.815595 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerStarted","Data":"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45"} Dec 12 07:08:38 crc kubenswrapper[4867]: I1212 07:08:38.818212 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerStarted","Data":"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee"} Dec 12 07:08:38 crc kubenswrapper[4867]: I1212 07:08:38.818267 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerStarted","Data":"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef"} Dec 12 07:08:38 crc kubenswrapper[4867]: I1212 07:08:38.842879 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.842857434 podStartE2EDuration="3.842857434s" podCreationTimestamp="2025-12-12 07:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:38.837985384 +0000 UTC m=+1206.409366653" watchObservedRunningTime="2025-12-12 07:08:38.842857434 +0000 UTC m=+1206.414238723" Dec 12 07:08:38 crc kubenswrapper[4867]: I1212 07:08:38.875045 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.875021626 podStartE2EDuration="2.875021626s" podCreationTimestamp="2025-12-12 07:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:38.86907009 +0000 UTC m=+1206.440451359" watchObservedRunningTime="2025-12-12 07:08:38.875021626 +0000 UTC m=+1206.446402895" Dec 12 07:08:39 crc kubenswrapper[4867]: I1212 07:08:39.790014 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:39 crc kubenswrapper[4867]: I1212 07:08:39.913462 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:42 crc kubenswrapper[4867]: I1212 07:08:42.220060 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.094451 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.157902 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.158442 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="dnsmasq-dns" containerID="cri-o://7046e803bfbfbbe7b0a68f19ce0ff19c35b971902172ac64fe4af0e5a0aff4bc" gracePeriod=10 Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.225718 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.345203 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.345424 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-579b864c64-lrpn7" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api-log" containerID="cri-o://a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e" gracePeriod=30 Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.345785 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-579b864c64-lrpn7" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" containerID="cri-o://461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50" gracePeriod=30 Dec 12 07:08:43 crc kubenswrapper[4867]: I1212 07:08:43.357142 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-579b864c64-lrpn7" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": EOF" Dec 12 07:08:44 crc kubenswrapper[4867]: I1212 07:08:44.877989 4867 generic.go:334] "Generic (PLEG): container finished" podID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerID="7046e803bfbfbbe7b0a68f19ce0ff19c35b971902172ac64fe4af0e5a0aff4bc" exitCode=0 Dec 12 07:08:44 crc kubenswrapper[4867]: I1212 07:08:44.878096 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" event={"ID":"bf3306f4-eec2-4188-b964-ac1cec81824a","Type":"ContainerDied","Data":"7046e803bfbfbbe7b0a68f19ce0ff19c35b971902172ac64fe4af0e5a0aff4bc"} Dec 12 07:08:44 crc kubenswrapper[4867]: I1212 07:08:44.880063 4867 generic.go:334] "Generic (PLEG): container finished" podID="ef816478-12ce-48f4-a83d-3800f623b879" containerID="a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e" exitCode=143 Dec 12 07:08:44 crc kubenswrapper[4867]: I1212 07:08:44.880124 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerDied","Data":"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e"} Dec 12 07:08:45 crc kubenswrapper[4867]: I1212 07:08:45.934134 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.303374 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.303703 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.345956 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.356973 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.434171 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.434219 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.468217 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.482666 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.753305 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lmsp" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.846803 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pck7x\" (UniqueName: \"kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x\") pod \"7299cd15-ce08-48f3-be09-c3a70901c4aa\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.846861 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs\") pod \"7299cd15-ce08-48f3-be09-c3a70901c4aa\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.846988 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data\") pod \"7299cd15-ce08-48f3-be09-c3a70901c4aa\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.847041 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle\") pod \"7299cd15-ce08-48f3-be09-c3a70901c4aa\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.847063 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts\") pod \"7299cd15-ce08-48f3-be09-c3a70901c4aa\" (UID: \"7299cd15-ce08-48f3-be09-c3a70901c4aa\") " Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.848217 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs" (OuterVolumeSpecName: "logs") pod "7299cd15-ce08-48f3-be09-c3a70901c4aa" (UID: "7299cd15-ce08-48f3-be09-c3a70901c4aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.874362 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts" (OuterVolumeSpecName: "scripts") pod "7299cd15-ce08-48f3-be09-c3a70901c4aa" (UID: "7299cd15-ce08-48f3-be09-c3a70901c4aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.883332 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data" (OuterVolumeSpecName: "config-data") pod "7299cd15-ce08-48f3-be09-c3a70901c4aa" (UID: "7299cd15-ce08-48f3-be09-c3a70901c4aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.888806 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x" (OuterVolumeSpecName: "kube-api-access-pck7x") pod "7299cd15-ce08-48f3-be09-c3a70901c4aa" (UID: "7299cd15-ce08-48f3-be09-c3a70901c4aa"). InnerVolumeSpecName "kube-api-access-pck7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.903396 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7299cd15-ce08-48f3-be09-c3a70901c4aa" (UID: "7299cd15-ce08-48f3-be09-c3a70901c4aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.912934 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6lmsp" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.949435 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pck7x\" (UniqueName: \"kubernetes.io/projected/7299cd15-ce08-48f3-be09-c3a70901c4aa-kube-api-access-pck7x\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.949468 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7299cd15-ce08-48f3-be09-c3a70901c4aa-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.949481 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.949491 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.949499 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7299cd15-ce08-48f3-be09-c3a70901c4aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984621 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984657 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6lmsp" event={"ID":"7299cd15-ce08-48f3-be09-c3a70901c4aa","Type":"ContainerDied","Data":"07e77352529ae62199afdaaa6bce28b46840142d0f0e38693a25ada31e6a71df"} Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984701 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07e77352529ae62199afdaaa6bce28b46840142d0f0e38693a25ada31e6a71df" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984719 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984731 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:46 crc kubenswrapper[4867]: I1212 07:08:46.984764 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.874820 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:08:47 crc kubenswrapper[4867]: E1212 07:08:47.875540 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" containerName="placement-db-sync" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.875562 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" containerName="placement-db-sync" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.875734 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" containerName="placement-db-sync" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.876656 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.880457 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.881093 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.881180 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.881444 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-hqhp8" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.889355 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.892460 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965573 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965623 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965680 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965751 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965767 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965834 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qph4s\" (UniqueName: \"kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:47 crc kubenswrapper[4867]: I1212 07:08:47.965880 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067205 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067349 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067376 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067437 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067515 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067537 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.067574 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qph4s\" (UniqueName: \"kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.068703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.075676 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.075790 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.077182 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.077699 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.092119 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.098053 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qph4s\" (UniqueName: \"kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s\") pod \"placement-68ccfcbfb4-q47vm\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.198651 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:48 crc kubenswrapper[4867]: E1212 07:08:48.669637 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c" Dec 12 07:08:48 crc kubenswrapper[4867]: E1212 07:08:48.669884 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmvlb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(de5ba6ce-674c-4887-8ba1-033e92e402c6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 12 07:08:48 crc kubenswrapper[4867]: E1212 07:08:48.671152 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.778884 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.882386 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.882794 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdnzk\" (UniqueName: \"kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.882869 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.883060 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.883198 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.883250 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0\") pod \"bf3306f4-eec2-4188-b964-ac1cec81824a\" (UID: \"bf3306f4-eec2-4188-b964-ac1cec81824a\") " Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.914457 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk" (OuterVolumeSpecName: "kube-api-access-sdnzk") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "kube-api-access-sdnzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.922211 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-579b864c64-lrpn7" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:49662->10.217.0.149:9311: read: connection reset by peer" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.922582 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-579b864c64-lrpn7" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:49668->10.217.0.149:9311: read: connection reset by peer" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.944797 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" event={"ID":"bf3306f4-eec2-4188-b964-ac1cec81824a","Type":"ContainerDied","Data":"dd837fc428c888b284c80683229a242eb4010f3b4aa4a44f78117cb707ad05c7"} Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.944868 4867 scope.go:117] "RemoveContainer" containerID="7046e803bfbfbbe7b0a68f19ce0ff19c35b971902172ac64fe4af0e5a0aff4bc" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.944923 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="ceilometer-notification-agent" containerID="cri-o://52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3" gracePeriod=30 Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945020 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945265 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945277 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945505 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="sg-core" containerID="cri-o://685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19" gracePeriod=30 Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945683 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.945709 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.985310 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdnzk\" (UniqueName: \"kubernetes.io/projected/bf3306f4-eec2-4188-b964-ac1cec81824a-kube-api-access-sdnzk\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.986700 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.990164 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:48 crc kubenswrapper[4867]: I1212 07:08:48.993476 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.001485 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config" (OuterVolumeSpecName: "config") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.001788 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf3306f4-eec2-4188-b964-ac1cec81824a" (UID: "bf3306f4-eec2-4188-b964-ac1cec81824a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.026759 4867 scope.go:117] "RemoveContainer" containerID="11d7633344c9a96316d54f1335579e380441baa68368ba0bd6c2d213c800734d" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.076780 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.087512 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.087546 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.087555 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.087564 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.087572 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf3306f4-eec2-4188-b964-ac1cec81824a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.172703 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.242283 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.269762 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.274103 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.420955 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.427922 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.437198 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fc6d4ffc7-8jtzn"] Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.493695 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs\") pod \"ef816478-12ce-48f4-a83d-3800f623b879\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.494054 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2sd8\" (UniqueName: \"kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8\") pod \"ef816478-12ce-48f4-a83d-3800f623b879\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.494160 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data\") pod \"ef816478-12ce-48f4-a83d-3800f623b879\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.494215 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom\") pod \"ef816478-12ce-48f4-a83d-3800f623b879\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.494314 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle\") pod \"ef816478-12ce-48f4-a83d-3800f623b879\" (UID: \"ef816478-12ce-48f4-a83d-3800f623b879\") " Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.496823 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs" (OuterVolumeSpecName: "logs") pod "ef816478-12ce-48f4-a83d-3800f623b879" (UID: "ef816478-12ce-48f4-a83d-3800f623b879"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.501575 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8" (OuterVolumeSpecName: "kube-api-access-n2sd8") pod "ef816478-12ce-48f4-a83d-3800f623b879" (UID: "ef816478-12ce-48f4-a83d-3800f623b879"). InnerVolumeSpecName "kube-api-access-n2sd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.513418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ef816478-12ce-48f4-a83d-3800f623b879" (UID: "ef816478-12ce-48f4-a83d-3800f623b879"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.526355 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef816478-12ce-48f4-a83d-3800f623b879" (UID: "ef816478-12ce-48f4-a83d-3800f623b879"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.563880 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data" (OuterVolumeSpecName: "config-data") pod "ef816478-12ce-48f4-a83d-3800f623b879" (UID: "ef816478-12ce-48f4-a83d-3800f623b879"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.596705 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.596744 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.596762 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef816478-12ce-48f4-a83d-3800f623b879-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.596774 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef816478-12ce-48f4-a83d-3800f623b879-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.596786 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2sd8\" (UniqueName: \"kubernetes.io/projected/ef816478-12ce-48f4-a83d-3800f623b879-kube-api-access-n2sd8\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.966134 4867 generic.go:334] "Generic (PLEG): container finished" podID="ef816478-12ce-48f4-a83d-3800f623b879" containerID="461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50" exitCode=0 Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.966302 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerDied","Data":"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50"} Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.966627 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-579b864c64-lrpn7" event={"ID":"ef816478-12ce-48f4-a83d-3800f623b879","Type":"ContainerDied","Data":"e698be70a47300a227af9224e30aa7c195bada2a74cb763728088013433ec5b9"} Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.966652 4867 scope.go:117] "RemoveContainer" containerID="461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.966414 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-579b864c64-lrpn7" Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.974215 4867 generic.go:334] "Generic (PLEG): container finished" podID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerID="685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19" exitCode=2 Dec 12 07:08:49 crc kubenswrapper[4867]: I1212 07:08:49.974304 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerDied","Data":"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19"} Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:49.983273 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerStarted","Data":"a6b387c7b3b91393496b3df6c1f76038666dc95195810690a8e0c383d5e2aa2d"} Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:49.983353 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerStarted","Data":"7d0afa62d13fd39fc3b3ba841aff538f37e9ec010eeac42f3c84ad166e59f590"} Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:49.983370 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerStarted","Data":"1a59081fd9e9b81afc0c2876cdea2a53be58c7cb530b3ea0ca07cb37fdd47e61"} Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:49.984421 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:49.984471 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.009619 4867 scope.go:117] "RemoveContainer" containerID="a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.013915 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.042106 4867 scope.go:117] "RemoveContainer" containerID="461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50" Dec 12 07:08:50 crc kubenswrapper[4867]: E1212 07:08:50.042714 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50\": container with ID starting with 461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50 not found: ID does not exist" containerID="461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.042774 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50"} err="failed to get container status \"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50\": rpc error: code = NotFound desc = could not find container \"461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50\": container with ID starting with 461cf2e53ddd3ed08fae27092a71a539568a464496c14c04c24ba4e296fc7f50 not found: ID does not exist" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.042810 4867 scope.go:117] "RemoveContainer" containerID="a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e" Dec 12 07:08:50 crc kubenswrapper[4867]: E1212 07:08:50.043344 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e\": container with ID starting with a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e not found: ID does not exist" containerID="a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.043423 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e"} err="failed to get container status \"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e\": rpc error: code = NotFound desc = could not find container \"a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e\": container with ID starting with a6cfa01d6fdf1b0a0645168b759f2e8d0753fccf6cf5a34637c7c10af205362e not found: ID does not exist" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.052019 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-579b864c64-lrpn7"] Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.056388 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68ccfcbfb4-q47vm" podStartSLOduration=3.056366321 podStartE2EDuration="3.056366321s" podCreationTimestamp="2025-12-12 07:08:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:50.044747355 +0000 UTC m=+1217.616128624" watchObservedRunningTime="2025-12-12 07:08:50.056366321 +0000 UTC m=+1217.627747590" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.850826 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" path="/var/lib/kubelet/pods/bf3306f4-eec2-4188-b964-ac1cec81824a/volumes" Dec 12 07:08:50 crc kubenswrapper[4867]: I1212 07:08:50.851997 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef816478-12ce-48f4-a83d-3800f623b879" path="/var/lib/kubelet/pods/ef816478-12ce-48f4-a83d-3800f623b879/volumes" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.634817 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.739892 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740162 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740220 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740312 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740508 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740554 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmvlb\" (UniqueName: \"kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740606 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml\") pod \"de5ba6ce-674c-4887-8ba1-033e92e402c6\" (UID: \"de5ba6ce-674c-4887-8ba1-033e92e402c6\") " Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.740320 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.741029 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.741355 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.741389 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de5ba6ce-674c-4887-8ba1-033e92e402c6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.748499 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts" (OuterVolumeSpecName: "scripts") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.748512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb" (OuterVolumeSpecName: "kube-api-access-kmvlb") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "kube-api-access-kmvlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.766150 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.767399 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.768687 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data" (OuterVolumeSpecName: "config-data") pod "de5ba6ce-674c-4887-8ba1-033e92e402c6" (UID: "de5ba6ce-674c-4887-8ba1-033e92e402c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.842575 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.842637 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.842648 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.842657 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de5ba6ce-674c-4887-8ba1-033e92e402c6-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:51 crc kubenswrapper[4867]: I1212 07:08:51.842665 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmvlb\" (UniqueName: \"kubernetes.io/projected/de5ba6ce-674c-4887-8ba1-033e92e402c6-kube-api-access-kmvlb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.003365 4867 generic.go:334] "Generic (PLEG): container finished" podID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerID="52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3" exitCode=0 Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.003435 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerDied","Data":"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3"} Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.003470 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de5ba6ce-674c-4887-8ba1-033e92e402c6","Type":"ContainerDied","Data":"05018dcd51b85f6cc97091f9c80d414dcf78b114274060274ccde5a89ff26c49"} Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.003490 4867 scope.go:117] "RemoveContainer" containerID="685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.003609 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.006574 4867 generic.go:334] "Generic (PLEG): container finished" podID="63341abf-5801-4239-8370-4c0732c95e3b" containerID="3f67093be35c467d200dbddfac2e3e8c6dc0d42da9b5d84d7d84b0c2b0ca891d" exitCode=0 Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.006633 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7fnd8" event={"ID":"63341abf-5801-4239-8370-4c0732c95e3b","Type":"ContainerDied","Data":"3f67093be35c467d200dbddfac2e3e8c6dc0d42da9b5d84d7d84b0c2b0ca891d"} Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.009838 4867 generic.go:334] "Generic (PLEG): container finished" podID="3f78a88a-c256-4e35-8f64-1c155650cca6" containerID="77646526d618a77ee8b6ba9c19111bdb9e0e386abe004e3170ce8918fdae8267" exitCode=0 Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.009871 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ptmqx" event={"ID":"3f78a88a-c256-4e35-8f64-1c155650cca6","Type":"ContainerDied","Data":"77646526d618a77ee8b6ba9c19111bdb9e0e386abe004e3170ce8918fdae8267"} Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.031839 4867 scope.go:117] "RemoveContainer" containerID="52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.057081 4867 scope.go:117] "RemoveContainer" containerID="685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.057615 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19\": container with ID starting with 685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19 not found: ID does not exist" containerID="685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.057657 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19"} err="failed to get container status \"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19\": rpc error: code = NotFound desc = could not find container \"685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19\": container with ID starting with 685666b688b6e54ffc2783067719f46fedad5b96cd5fd653324a40dec45b2f19 not found: ID does not exist" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.057684 4867 scope.go:117] "RemoveContainer" containerID="52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.057947 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3\": container with ID starting with 52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3 not found: ID does not exist" containerID="52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.057979 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3"} err="failed to get container status \"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3\": rpc error: code = NotFound desc = could not find container \"52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3\": container with ID starting with 52d614242eee0639d56a655b013018636d21a1e3913fe185ba7c634ab17eecc3 not found: ID does not exist" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.100400 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.110746 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116069 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116507 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="init" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116527 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="init" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116548 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116556 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116575 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="ceilometer-notification-agent" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116582 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="ceilometer-notification-agent" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116600 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="dnsmasq-dns" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116607 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="dnsmasq-dns" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116626 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api-log" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116632 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api-log" Dec 12 07:08:52 crc kubenswrapper[4867]: E1212 07:08:52.116647 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="sg-core" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116653 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="sg-core" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116812 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api-log" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116826 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf3306f4-eec2-4188-b964-ac1cec81824a" containerName="dnsmasq-dns" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116841 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="sg-core" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116854 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" containerName="ceilometer-notification-agent" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.116862 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef816478-12ce-48f4-a83d-3800f623b879" containerName="barbican-api" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.118951 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.124061 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.124313 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.124532 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.249801 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250176 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250210 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250306 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250363 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqr74\" (UniqueName: \"kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250493 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.250516 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.351885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.351937 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.351957 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352010 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352045 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqr74\" (UniqueName: \"kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352132 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352148 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352468 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.352473 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.356117 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.356753 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.362032 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.362807 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.374842 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqr74\" (UniqueName: \"kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74\") pod \"ceilometer-0\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.439391 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.853898 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de5ba6ce-674c-4887-8ba1-033e92e402c6" path="/var/lib/kubelet/pods/de5ba6ce-674c-4887-8ba1-033e92e402c6/volumes" Dec 12 07:08:52 crc kubenswrapper[4867]: I1212 07:08:52.858191 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.020872 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerStarted","Data":"a2c54ce60703c602d9d0a200e38c2d9ab3e3f3bae26846b054c60e018e1c1b4e"} Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.306409 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.371484 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle\") pod \"3f78a88a-c256-4e35-8f64-1c155650cca6\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.371603 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config\") pod \"3f78a88a-c256-4e35-8f64-1c155650cca6\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.371829 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67j8q\" (UniqueName: \"kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q\") pod \"3f78a88a-c256-4e35-8f64-1c155650cca6\" (UID: \"3f78a88a-c256-4e35-8f64-1c155650cca6\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.380197 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q" (OuterVolumeSpecName: "kube-api-access-67j8q") pod "3f78a88a-c256-4e35-8f64-1c155650cca6" (UID: "3f78a88a-c256-4e35-8f64-1c155650cca6"). InnerVolumeSpecName "kube-api-access-67j8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.400889 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config" (OuterVolumeSpecName: "config") pod "3f78a88a-c256-4e35-8f64-1c155650cca6" (UID: "3f78a88a-c256-4e35-8f64-1c155650cca6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.404027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f78a88a-c256-4e35-8f64-1c155650cca6" (UID: "3f78a88a-c256-4e35-8f64-1c155650cca6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.474619 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.474666 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67j8q\" (UniqueName: \"kubernetes.io/projected/3f78a88a-c256-4e35-8f64-1c155650cca6-kube-api-access-67j8q\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.474681 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f78a88a-c256-4e35-8f64-1c155650cca6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.490570 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.576559 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42v2r\" (UniqueName: \"kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.576643 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.576672 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.576798 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.576957 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.577144 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.577506 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data\") pod \"63341abf-5801-4239-8370-4c0732c95e3b\" (UID: \"63341abf-5801-4239-8370-4c0732c95e3b\") " Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.578116 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/63341abf-5801-4239-8370-4c0732c95e3b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.582783 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.583365 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r" (OuterVolumeSpecName: "kube-api-access-42v2r") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "kube-api-access-42v2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.588209 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts" (OuterVolumeSpecName: "scripts") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.610268 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.648000 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data" (OuterVolumeSpecName: "config-data") pod "63341abf-5801-4239-8370-4c0732c95e3b" (UID: "63341abf-5801-4239-8370-4c0732c95e3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.679663 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.679706 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.679721 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42v2r\" (UniqueName: \"kubernetes.io/projected/63341abf-5801-4239-8370-4c0732c95e3b-kube-api-access-42v2r\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.679734 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:53 crc kubenswrapper[4867]: I1212 07:08:53.679743 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63341abf-5801-4239-8370-4c0732c95e3b-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.035936 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7fnd8" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.035958 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7fnd8" event={"ID":"63341abf-5801-4239-8370-4c0732c95e3b","Type":"ContainerDied","Data":"c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1"} Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.036006 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c29bb2dc402a7388ecfd0df54a0cca78d5b87db3f87987f0fb8f51ac56e003e1" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.037661 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ptmqx" event={"ID":"3f78a88a-c256-4e35-8f64-1c155650cca6","Type":"ContainerDied","Data":"dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235"} Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.037685 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd16b9c07320c2626789a3970ae5a91ec9c233e4c5ba9d7a273cdd0b28a71235" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.037736 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ptmqx" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.299062 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66cdd4b5b5-f5rg2"] Dec 12 07:08:54 crc kubenswrapper[4867]: E1212 07:08:54.299562 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63341abf-5801-4239-8370-4c0732c95e3b" containerName="cinder-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.299596 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="63341abf-5801-4239-8370-4c0732c95e3b" containerName="cinder-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: E1212 07:08:54.299626 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f78a88a-c256-4e35-8f64-1c155650cca6" containerName="neutron-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.299636 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f78a88a-c256-4e35-8f64-1c155650cca6" containerName="neutron-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.299852 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f78a88a-c256-4e35-8f64-1c155650cca6" containerName="neutron-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.299878 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="63341abf-5801-4239-8370-4c0732c95e3b" containerName="cinder-db-sync" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.304016 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.324947 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cdd4b5b5-f5rg2"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.395571 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwptw\" (UniqueName: \"kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.395940 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.396010 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.396052 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.396086 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.396116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.412341 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.414200 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.419666 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-r2vqq" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.419903 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.420597 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.420737 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.453961 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499145 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499239 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499322 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499356 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh7b5\" (UniqueName: \"kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499411 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499490 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499517 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499564 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499582 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499599 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwptw\" (UniqueName: \"kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.499617 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.500047 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cdd4b5b5-f5rg2"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.500516 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.500649 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: E1212 07:08:54.500694 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-dwptw ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" podUID="44839143-2149-4ced-bbbd-2c855155c934" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.501212 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.505103 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.506392 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.533655 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.535468 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.539374 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.539431 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-k65fc" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.539592 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.539667 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.573243 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwptw\" (UniqueName: \"kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw\") pod \"dnsmasq-dns-66cdd4b5b5-f5rg2\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.574953 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.593909 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.595569 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.609907 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.619716 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620045 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620194 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh7b5\" (UniqueName: \"kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620337 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620487 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620650 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620731 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lskb\" (UniqueName: \"kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620874 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.620942 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.621018 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.621098 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.609977 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.633918 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.633988 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.641019 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.652984 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.678445 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh7b5\" (UniqueName: \"kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.689472 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data\") pod \"cinder-scheduler-0\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.699314 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.701390 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.709008 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.716775 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727184 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727263 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lskb\" (UniqueName: \"kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727300 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727331 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f9wk\" (UniqueName: \"kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727368 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727445 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727740 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727771 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727818 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727844 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727868 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727896 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727920 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727939 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727958 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.727990 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh8h5\" (UniqueName: \"kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.728958 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.729670 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.730404 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.730736 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.742643 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.765814 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.770263 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lskb\" (UniqueName: \"kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb\") pod \"dnsmasq-dns-75dbb546bf-dqprf\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841151 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841214 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841464 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh8h5\" (UniqueName: \"kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841792 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841852 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f9wk\" (UniqueName: \"kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841896 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.841990 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.842139 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.842162 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.842197 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.842242 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.842298 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.847109 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.848035 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.850032 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.857555 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.867600 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.867722 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.868313 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.868889 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.869686 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.886093 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.911506 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f9wk\" (UniqueName: \"kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk\") pod \"neutron-6c7bdbbdf4-sbn95\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:54 crc kubenswrapper[4867]: I1212 07:08:54.935105 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh8h5\" (UniqueName: \"kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5\") pod \"cinder-api-0\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " pod="openstack/cinder-api-0" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.035785 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.064973 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.089705 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.090412 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerStarted","Data":"f0e95653ca928d0a0d0624cd8eace10ce6e6530062f0230018c62dd0396df2b4"} Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.107859 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.131860 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.252571 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.252709 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.253893 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.254197 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.256067 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config" (OuterVolumeSpecName: "config") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.255331 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.258610 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwptw\" (UniqueName: \"kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.258771 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.258826 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0\") pod \"44839143-2149-4ced-bbbd-2c855155c934\" (UID: \"44839143-2149-4ced-bbbd-2c855155c934\") " Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.259912 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.260961 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.260983 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.260998 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.261010 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.264499 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.279686 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw" (OuterVolumeSpecName: "kube-api-access-dwptw") pod "44839143-2149-4ced-bbbd-2c855155c934" (UID: "44839143-2149-4ced-bbbd-2c855155c934"). InnerVolumeSpecName "kube-api-access-dwptw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.362773 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwptw\" (UniqueName: \"kubernetes.io/projected/44839143-2149-4ced-bbbd-2c855155c934-kube-api-access-dwptw\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.362811 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/44839143-2149-4ced-bbbd-2c855155c934-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.392671 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:08:55 crc kubenswrapper[4867]: W1212 07:08:55.422791 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e36f8ad_95b1_454d_bdc0_6812957e7b1f.slice/crio-de0e17d27cd318e6473aac4cade957ed88785b97b31d30433e66f19234c7e55c WatchSource:0}: Error finding container de0e17d27cd318e6473aac4cade957ed88785b97b31d30433e66f19234c7e55c: Status 404 returned error can't find the container with id de0e17d27cd318e6473aac4cade957ed88785b97b31d30433e66f19234c7e55c Dec 12 07:08:55 crc kubenswrapper[4867]: W1212 07:08:55.770264 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e6579f1_6cee_40dc_b1b1_fa309701574d.slice/crio-10e261a2f434f43d7b0e451c7b4e80c28dea9a8eb57c9abd66a691ac2b54efb9 WatchSource:0}: Error finding container 10e261a2f434f43d7b0e451c7b4e80c28dea9a8eb57c9abd66a691ac2b54efb9: Status 404 returned error can't find the container with id 10e261a2f434f43d7b0e451c7b4e80c28dea9a8eb57c9abd66a691ac2b54efb9 Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.773542 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.820477 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:08:55 crc kubenswrapper[4867]: I1212 07:08:55.871311 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.104305 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerStarted","Data":"c7b6983972876b0fbbdb417fdbf7b381d8d49b5a245c8760ab6ca7e9b75635e6"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.116236 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerStarted","Data":"518289223b7438ceaba6563b966ae4a359e3859806f0fefbccb32ce3766718f2"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.125103 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerStarted","Data":"c811c84680ea4faba42aafce3b8064b9916234e79bc5afdfd21586aad101aaad"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.143515 4867 generic.go:334] "Generic (PLEG): container finished" podID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerID="1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6" exitCode=0 Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.143657 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" event={"ID":"6e6579f1-6cee-40dc-b1b1-fa309701574d","Type":"ContainerDied","Data":"1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.143735 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" event={"ID":"6e6579f1-6cee-40dc-b1b1-fa309701574d","Type":"ContainerStarted","Data":"10e261a2f434f43d7b0e451c7b4e80c28dea9a8eb57c9abd66a691ac2b54efb9"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.158829 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cdd4b5b5-f5rg2" Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.159137 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerStarted","Data":"de0e17d27cd318e6473aac4cade957ed88785b97b31d30433e66f19234c7e55c"} Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.258040 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cdd4b5b5-f5rg2"] Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.259561 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66cdd4b5b5-f5rg2"] Dec 12 07:08:56 crc kubenswrapper[4867]: I1212 07:08:56.854913 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44839143-2149-4ced-bbbd-2c855155c934" path="/var/lib/kubelet/pods/44839143-2149-4ced-bbbd-2c855155c934/volumes" Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.184977 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerStarted","Data":"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c"} Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.189405 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerStarted","Data":"4c1fdd98d9ae557027eb96d541bae60198656d3677292068ba260ca74e4959e1"} Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.191314 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerStarted","Data":"4f9fc24b43d1f00ce0c09640b3c2ab26e264143888b3092794852d66e2f6fd8b"} Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.191349 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerStarted","Data":"5ee0aa180aea7ab1ad6875c0ea5b1b61058a02dcd72288d4cf0df2058299c394"} Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.192821 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.202061 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" event={"ID":"6e6579f1-6cee-40dc-b1b1-fa309701574d","Type":"ContainerStarted","Data":"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc"} Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.202384 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.224784 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c7bdbbdf4-sbn95" podStartSLOduration=3.224753117 podStartE2EDuration="3.224753117s" podCreationTimestamp="2025-12-12 07:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:57.221122477 +0000 UTC m=+1224.792503766" watchObservedRunningTime="2025-12-12 07:08:57.224753117 +0000 UTC m=+1224.796134386" Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.264015 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:57 crc kubenswrapper[4867]: I1212 07:08:57.272673 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" podStartSLOduration=3.272652656 podStartE2EDuration="3.272652656s" podCreationTimestamp="2025-12-12 07:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:57.264451194 +0000 UTC m=+1224.835832483" watchObservedRunningTime="2025-12-12 07:08:57.272652656 +0000 UTC m=+1224.844033925" Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.213992 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerStarted","Data":"1ab5f33993cfd234d44451590b9813416e61710c093abb6ba1ae3ea0663d1f29"} Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.218884 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerStarted","Data":"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166"} Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.219067 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api-log" containerID="cri-o://4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" gracePeriod=30 Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.219374 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.219713 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api" containerID="cri-o://f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" gracePeriod=30 Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.229804 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerStarted","Data":"ef3d627b3369c94b9be6839b951980966127172afc75b31b3abfa1f3a3fb0fed"} Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.230158 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.243458 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.243435637 podStartE2EDuration="4.243435637s" podCreationTimestamp="2025-12-12 07:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:08:58.238747732 +0000 UTC m=+1225.810129001" watchObservedRunningTime="2025-12-12 07:08:58.243435637 +0000 UTC m=+1225.814816916" Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.281961 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.471981884 podStartE2EDuration="6.281939485s" podCreationTimestamp="2025-12-12 07:08:52 +0000 UTC" firstStartedPulling="2025-12-12 07:08:52.864365005 +0000 UTC m=+1220.435746274" lastFinishedPulling="2025-12-12 07:08:57.674322606 +0000 UTC m=+1225.245703875" observedRunningTime="2025-12-12 07:08:58.274347139 +0000 UTC m=+1225.845728408" watchObservedRunningTime="2025-12-12 07:08:58.281939485 +0000 UTC m=+1225.853320754" Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.989771 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:08:58 crc kubenswrapper[4867]: I1212 07:08:58.990314 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.130850 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241555 4867 generic.go:334] "Generic (PLEG): container finished" podID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerID="f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" exitCode=0 Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241594 4867 generic.go:334] "Generic (PLEG): container finished" podID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerID="4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" exitCode=143 Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241604 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241653 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerDied","Data":"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166"} Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241686 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerDied","Data":"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c"} Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241699 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"182b01de-55c1-4d41-b8fe-387ba97b1b28","Type":"ContainerDied","Data":"c7b6983972876b0fbbdb417fdbf7b381d8d49b5a245c8760ab6ca7e9b75635e6"} Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.241719 4867 scope.go:117] "RemoveContainer" containerID="f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.247809 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerStarted","Data":"f6ba841c180eb95a59980756860025113355d437a8b64663f1e109cf34cd4ecc"} Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258013 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh8h5\" (UniqueName: \"kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258106 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258169 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258330 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258409 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258486 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.258548 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom\") pod \"182b01de-55c1-4d41-b8fe-387ba97b1b28\" (UID: \"182b01de-55c1-4d41-b8fe-387ba97b1b28\") " Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.259413 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.259427 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs" (OuterVolumeSpecName: "logs") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.259656 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/182b01de-55c1-4d41-b8fe-387ba97b1b28-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.259701 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/182b01de-55c1-4d41-b8fe-387ba97b1b28-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.265891 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.267855 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5" (OuterVolumeSpecName: "kube-api-access-hh8h5") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "kube-api-access-hh8h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.277696 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts" (OuterVolumeSpecName: "scripts") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.279707 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.121938396 podStartE2EDuration="5.27968309s" podCreationTimestamp="2025-12-12 07:08:54 +0000 UTC" firstStartedPulling="2025-12-12 07:08:55.424804112 +0000 UTC m=+1222.996185391" lastFinishedPulling="2025-12-12 07:08:56.582548816 +0000 UTC m=+1224.153930085" observedRunningTime="2025-12-12 07:08:59.270901693 +0000 UTC m=+1226.842282982" watchObservedRunningTime="2025-12-12 07:08:59.27968309 +0000 UTC m=+1226.851064369" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.282471 4867 scope.go:117] "RemoveContainer" containerID="4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.306992 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.339446 4867 scope.go:117] "RemoveContainer" containerID="f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" Dec 12 07:08:59 crc kubenswrapper[4867]: E1212 07:08:59.339958 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166\": container with ID starting with f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166 not found: ID does not exist" containerID="f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.340095 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166"} err="failed to get container status \"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166\": rpc error: code = NotFound desc = could not find container \"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166\": container with ID starting with f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166 not found: ID does not exist" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.340189 4867 scope.go:117] "RemoveContainer" containerID="4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" Dec 12 07:08:59 crc kubenswrapper[4867]: E1212 07:08:59.340535 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c\": container with ID starting with 4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c not found: ID does not exist" containerID="4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.340630 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c"} err="failed to get container status \"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c\": rpc error: code = NotFound desc = could not find container \"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c\": container with ID starting with 4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c not found: ID does not exist" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.340709 4867 scope.go:117] "RemoveContainer" containerID="f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.341098 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166"} err="failed to get container status \"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166\": rpc error: code = NotFound desc = could not find container \"f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166\": container with ID starting with f0951d9d3691f26cb39a77821b44c1d0dd2a7987f8d6ed037c6b08b7bb1ac166 not found: ID does not exist" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.341195 4867 scope.go:117] "RemoveContainer" containerID="4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.341552 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c"} err="failed to get container status \"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c\": rpc error: code = NotFound desc = could not find container \"4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c\": container with ID starting with 4735a4e5d4d493b2ec857c6aee6cc60c4b676871baba47bebf78a6be4376d30c not found: ID does not exist" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.364186 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.364420 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.364501 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh8h5\" (UniqueName: \"kubernetes.io/projected/182b01de-55c1-4d41-b8fe-387ba97b1b28-kube-api-access-hh8h5\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.364569 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.375287 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data" (OuterVolumeSpecName: "config-data") pod "182b01de-55c1-4d41-b8fe-387ba97b1b28" (UID: "182b01de-55c1-4d41-b8fe-387ba97b1b28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.466837 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/182b01de-55c1-4d41-b8fe-387ba97b1b28-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.574623 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.587288 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.597275 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:59 crc kubenswrapper[4867]: E1212 07:08:59.597733 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api-log" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.597756 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api-log" Dec 12 07:08:59 crc kubenswrapper[4867]: E1212 07:08:59.597790 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.597798 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.598036 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.598075 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" containerName="cinder-api-log" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.599327 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.606674 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.607737 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.607944 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.625353 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.766962 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773644 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773724 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773760 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773781 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.773974 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.774162 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjt7k\" (UniqueName: \"kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.774241 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.774282 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.875750 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.875887 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjt7k\" (UniqueName: \"kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.875942 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.875978 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876038 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876068 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876123 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876179 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876217 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.876754 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.880480 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.880579 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.881298 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.882623 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.883962 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.883970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.884686 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.897506 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjt7k\" (UniqueName: \"kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k\") pod \"cinder-api-0\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " pod="openstack/cinder-api-0" Dec 12 07:08:59 crc kubenswrapper[4867]: I1212 07:08:59.933971 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:09:00 crc kubenswrapper[4867]: I1212 07:09:00.437540 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:09:00 crc kubenswrapper[4867]: I1212 07:09:00.849496 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="182b01de-55c1-4d41-b8fe-387ba97b1b28" path="/var/lib/kubelet/pods/182b01de-55c1-4d41-b8fe-387ba97b1b28/volumes" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.275689 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerStarted","Data":"f3626d4c1ef6132f0076b910eab89601c61a779670954414124b9e2ed04cba99"} Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.275996 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerStarted","Data":"8e9487be27e4f37fb16b1e5172bed50b567bef99991375dcb870b256497c8730"} Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.478503 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.480403 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.486744 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.489248 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.494840 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610513 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610578 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610594 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mb7\" (UniqueName: \"kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610712 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.610975 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.611079 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.713715 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.713838 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.713886 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mb7\" (UniqueName: \"kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.713914 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.714040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.714078 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.714161 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.725518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.726299 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.734438 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.736639 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mb7\" (UniqueName: \"kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.737046 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.743094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.758562 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config\") pod \"neutron-6796c56757-d2snl\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:01 crc kubenswrapper[4867]: I1212 07:09:01.802812 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:02 crc kubenswrapper[4867]: I1212 07:09:02.233729 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:09:02 crc kubenswrapper[4867]: I1212 07:09:02.490835 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.310512 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerStarted","Data":"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478"} Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.311130 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerStarted","Data":"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b"} Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.311152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerStarted","Data":"6977f05a7db0225dce26c456a746ae79bf41e55e65070e03c867704eba3d3f00"} Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.311204 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.317414 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerStarted","Data":"384bb1fccf6418383d28e93ad17cb3f3d26be31108864c9e588482c7982d8d1c"} Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.317709 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.340360 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6796c56757-d2snl" podStartSLOduration=2.340342553 podStartE2EDuration="2.340342553s" podCreationTimestamp="2025-12-12 07:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:03.338854557 +0000 UTC m=+1230.910235826" watchObservedRunningTime="2025-12-12 07:09:03.340342553 +0000 UTC m=+1230.911723822" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.402823 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.402806421 podStartE2EDuration="4.402806421s" podCreationTimestamp="2025-12-12 07:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:03.379945008 +0000 UTC m=+1230.951326297" watchObservedRunningTime="2025-12-12 07:09:03.402806421 +0000 UTC m=+1230.974187690" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.466768 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.469081 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.473673 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.474126 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.474326 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xkrtt" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.497626 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.575027 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.575335 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5pbb\" (UniqueName: \"kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.575629 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.575695 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.677198 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.678030 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5pbb\" (UniqueName: \"kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.678279 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.678391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.679638 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.682591 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.683307 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.701804 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5pbb\" (UniqueName: \"kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb\") pod \"openstackclient\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " pod="openstack/openstackclient" Dec 12 07:09:03 crc kubenswrapper[4867]: I1212 07:09:03.822771 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 07:09:04 crc kubenswrapper[4867]: I1212 07:09:04.343489 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 07:09:04 crc kubenswrapper[4867]: W1212 07:09:04.349851 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod479f0ca7_1408_41d4_abaf_6034d0802100.slice/crio-2bf83ff8ff1761b6befca87c9d4fe5744f41cb77ca3600d5d550b6f39b1a9480 WatchSource:0}: Error finding container 2bf83ff8ff1761b6befca87c9d4fe5744f41cb77ca3600d5d550b6f39b1a9480: Status 404 returned error can't find the container with id 2bf83ff8ff1761b6befca87c9d4fe5744f41cb77ca3600d5d550b6f39b1a9480 Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.023720 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.067418 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.093788 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.150895 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.151335 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="dnsmasq-dns" containerID="cri-o://86e3b995214e4f591f726252c55ce30449f93230f261366531b28c1a122cd8c5" gracePeriod=10 Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.355263 4867 generic.go:334] "Generic (PLEG): container finished" podID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerID="86e3b995214e4f591f726252c55ce30449f93230f261366531b28c1a122cd8c5" exitCode=0 Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.355362 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" event={"ID":"03a0f30d-fe35-4e6b-bdfd-746f1040c82b","Type":"ContainerDied","Data":"86e3b995214e4f591f726252c55ce30449f93230f261366531b28c1a122cd8c5"} Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.358551 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"479f0ca7-1408-41d4-abaf-6034d0802100","Type":"ContainerStarted","Data":"2bf83ff8ff1761b6befca87c9d4fe5744f41cb77ca3600d5d550b6f39b1a9480"} Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.358704 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="cinder-scheduler" containerID="cri-o://1ab5f33993cfd234d44451590b9813416e61710c093abb6ba1ae3ea0663d1f29" gracePeriod=30 Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.359218 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="probe" containerID="cri-o://f6ba841c180eb95a59980756860025113355d437a8b64663f1e109cf34cd4ecc" gracePeriod=30 Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.744688 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922403 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922530 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26h6f\" (UniqueName: \"kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922570 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922650 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922740 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.922770 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0\") pod \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\" (UID: \"03a0f30d-fe35-4e6b-bdfd-746f1040c82b\") " Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.938680 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f" (OuterVolumeSpecName: "kube-api-access-26h6f") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "kube-api-access-26h6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.976617 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:05 crc kubenswrapper[4867]: I1212 07:09:05.980912 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:05.996892 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.004640 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.022258 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config" (OuterVolumeSpecName: "config") pod "03a0f30d-fe35-4e6b-bdfd-746f1040c82b" (UID: "03a0f30d-fe35-4e6b-bdfd-746f1040c82b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025784 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025825 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025839 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025851 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26h6f\" (UniqueName: \"kubernetes.io/projected/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-kube-api-access-26h6f\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025861 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.025871 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/03a0f30d-fe35-4e6b-bdfd-746f1040c82b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.373602 4867 generic.go:334] "Generic (PLEG): container finished" podID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerID="f6ba841c180eb95a59980756860025113355d437a8b64663f1e109cf34cd4ecc" exitCode=0 Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.373684 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerDied","Data":"f6ba841c180eb95a59980756860025113355d437a8b64663f1e109cf34cd4ecc"} Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.375662 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" event={"ID":"03a0f30d-fe35-4e6b-bdfd-746f1040c82b","Type":"ContainerDied","Data":"8f579aef811a9e4d9a2e7f3f30be0188f7311eb730690bced8c1cb5f44eb33e9"} Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.375767 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fffc8985-2v9wn" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.375818 4867 scope.go:117] "RemoveContainer" containerID="86e3b995214e4f591f726252c55ce30449f93230f261366531b28c1a122cd8c5" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.415419 4867 scope.go:117] "RemoveContainer" containerID="7608d1d79d129f76784e50627ed23dd3a9f295ec2ef5e27466b66734c863a038" Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.418340 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.430534 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8fffc8985-2v9wn"] Dec 12 07:09:06 crc kubenswrapper[4867]: I1212 07:09:06.854048 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" path="/var/lib/kubelet/pods/03a0f30d-fe35-4e6b-bdfd-746f1040c82b/volumes" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.864074 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:09:07 crc kubenswrapper[4867]: E1212 07:09:07.864806 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="dnsmasq-dns" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.864824 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="dnsmasq-dns" Dec 12 07:09:07 crc kubenswrapper[4867]: E1212 07:09:07.864844 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="init" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.864852 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="init" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.865097 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a0f30d-fe35-4e6b-bdfd-746f1040c82b" containerName="dnsmasq-dns" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.866178 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.868836 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.869434 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.872721 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.911582 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969383 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969632 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969744 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969825 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qfbb\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:07 crc kubenswrapper[4867]: I1212 07:09:07.969906 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.071545 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.071604 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.071693 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.072341 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.072548 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.073064 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.073125 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.073155 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qfbb\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.073259 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.073489 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.080767 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.083652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.084551 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.084598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.087588 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.092140 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qfbb\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb\") pod \"swift-proxy-67c5f5d445-rd5dx\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.199654 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.865544 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.954851 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.955241 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-central-agent" containerID="cri-o://f0e95653ca928d0a0d0624cd8eace10ce6e6530062f0230018c62dd0396df2b4" gracePeriod=30 Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.955395 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="proxy-httpd" containerID="cri-o://ef3d627b3369c94b9be6839b951980966127172afc75b31b3abfa1f3a3fb0fed" gracePeriod=30 Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.955456 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="sg-core" containerID="cri-o://4c1fdd98d9ae557027eb96d541bae60198656d3677292068ba260ca74e4959e1" gracePeriod=30 Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.955515 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-notification-agent" containerID="cri-o://518289223b7438ceaba6563b966ae4a359e3859806f0fefbccb32ce3766718f2" gracePeriod=30 Dec 12 07:09:08 crc kubenswrapper[4867]: I1212 07:09:08.968892 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.155:3000/\": EOF" Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429296 4867 generic.go:334] "Generic (PLEG): container finished" podID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerID="ef3d627b3369c94b9be6839b951980966127172afc75b31b3abfa1f3a3fb0fed" exitCode=0 Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429535 4867 generic.go:334] "Generic (PLEG): container finished" podID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerID="4c1fdd98d9ae557027eb96d541bae60198656d3677292068ba260ca74e4959e1" exitCode=2 Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429547 4867 generic.go:334] "Generic (PLEG): container finished" podID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerID="f0e95653ca928d0a0d0624cd8eace10ce6e6530062f0230018c62dd0396df2b4" exitCode=0 Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429403 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerDied","Data":"ef3d627b3369c94b9be6839b951980966127172afc75b31b3abfa1f3a3fb0fed"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429610 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerDied","Data":"4c1fdd98d9ae557027eb96d541bae60198656d3677292068ba260ca74e4959e1"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.429625 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerDied","Data":"f0e95653ca928d0a0d0624cd8eace10ce6e6530062f0230018c62dd0396df2b4"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.433085 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerStarted","Data":"bc30a378f2f47b54d0e25ab70eab05c43f7eb5652e82cd1c2e5b19b297703f21"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.433138 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerStarted","Data":"b4cea50b4a2409e5c06031d89001030cb96b03f21678fb75ce73c9b0ea30ae49"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.435157 4867 generic.go:334] "Generic (PLEG): container finished" podID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerID="1ab5f33993cfd234d44451590b9813416e61710c093abb6ba1ae3ea0663d1f29" exitCode=0 Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.435202 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerDied","Data":"1ab5f33993cfd234d44451590b9813416e61710c093abb6ba1ae3ea0663d1f29"} Dec 12 07:09:09 crc kubenswrapper[4867]: I1212 07:09:09.921468 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012067 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012185 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012273 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012391 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh7b5\" (UniqueName: \"kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012472 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012528 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom\") pod \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\" (UID: \"8e36f8ad-95b1-454d-bdc0-6812957e7b1f\") " Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.012378 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.013051 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.018254 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5" (OuterVolumeSpecName: "kube-api-access-wh7b5") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "kube-api-access-wh7b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.018874 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts" (OuterVolumeSpecName: "scripts") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.022446 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.114720 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh7b5\" (UniqueName: \"kubernetes.io/projected/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-kube-api-access-wh7b5\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.114749 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.114760 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.169835 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.214581 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data" (OuterVolumeSpecName: "config-data") pod "8e36f8ad-95b1-454d-bdc0-6812957e7b1f" (UID: "8e36f8ad-95b1-454d-bdc0-6812957e7b1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.217757 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.217792 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e36f8ad-95b1-454d-bdc0-6812957e7b1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.444850 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerStarted","Data":"dc60c646092d5fd6e48768335070dbf4389a632716a9616ea70be7334134e290"} Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.445591 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.446039 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.448952 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8e36f8ad-95b1-454d-bdc0-6812957e7b1f","Type":"ContainerDied","Data":"de0e17d27cd318e6473aac4cade957ed88785b97b31d30433e66f19234c7e55c"} Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.448984 4867 scope.go:117] "RemoveContainer" containerID="f6ba841c180eb95a59980756860025113355d437a8b64663f1e109cf34cd4ecc" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.449115 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.475931 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-67c5f5d445-rd5dx" podStartSLOduration=3.475913937 podStartE2EDuration="3.475913937s" podCreationTimestamp="2025-12-12 07:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:10.47562207 +0000 UTC m=+1238.047003359" watchObservedRunningTime="2025-12-12 07:09:10.475913937 +0000 UTC m=+1238.047295206" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.494185 4867 scope.go:117] "RemoveContainer" containerID="1ab5f33993cfd234d44451590b9813416e61710c093abb6ba1ae3ea0663d1f29" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.513181 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.521855 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.538092 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:10 crc kubenswrapper[4867]: E1212 07:09:10.538838 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="probe" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.538871 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="probe" Dec 12 07:09:10 crc kubenswrapper[4867]: E1212 07:09:10.538894 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="cinder-scheduler" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.538902 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="cinder-scheduler" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.539168 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="probe" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.539190 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" containerName="cinder-scheduler" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.556925 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.559781 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.594733 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.628557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.628751 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.628893 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.629077 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.629216 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppg6d\" (UniqueName: \"kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.629587 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733023 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppg6d\" (UniqueName: \"kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733258 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733305 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733417 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733431 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733511 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.733613 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.738906 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.745288 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.747449 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.756166 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.770655 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppg6d\" (UniqueName: \"kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d\") pod \"cinder-scheduler-0\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " pod="openstack/cinder-scheduler-0" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.851629 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e36f8ad-95b1-454d-bdc0-6812957e7b1f" path="/var/lib/kubelet/pods/8e36f8ad-95b1-454d-bdc0-6812957e7b1f/volumes" Dec 12 07:09:10 crc kubenswrapper[4867]: I1212 07:09:10.907175 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:09:12 crc kubenswrapper[4867]: I1212 07:09:12.276318 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 12 07:09:13 crc kubenswrapper[4867]: I1212 07:09:13.486789 4867 generic.go:334] "Generic (PLEG): container finished" podID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerID="518289223b7438ceaba6563b966ae4a359e3859806f0fefbccb32ce3766718f2" exitCode=0 Dec 12 07:09:13 crc kubenswrapper[4867]: I1212 07:09:13.486877 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerDied","Data":"518289223b7438ceaba6563b966ae4a359e3859806f0fefbccb32ce3766718f2"} Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.148821 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-rchnq"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.150091 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.157884 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rchnq"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.204004 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.204097 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cswv5\" (UniqueName: \"kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.263642 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tc5wq"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.265328 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.291110 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tc5wq"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.307547 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.307660 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cswv5\" (UniqueName: \"kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.308671 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.359130 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cswv5\" (UniqueName: \"kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5\") pod \"nova-api-db-create-rchnq\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.381484 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b624l"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.382843 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.406373 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-da12-account-create-update-wttp8"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.409995 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.420837 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.421687 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b624l"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.436314 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.436490 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb4nj\" (UniqueName: \"kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.460212 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-da12-account-create-update-wttp8"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.466547 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540448 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x6n6\" (UniqueName: \"kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540535 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540571 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540753 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540867 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb4nj\" (UniqueName: \"kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.540897 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zz47\" (UniqueName: \"kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.541336 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.556781 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-896e-account-create-update-x9tvp"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.558193 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.563652 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.570967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb4nj\" (UniqueName: \"kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj\") pod \"nova-cell0-db-create-tc5wq\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.573572 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-896e-account-create-update-x9tvp"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.591717 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643246 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x6n6\" (UniqueName: \"kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643361 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643404 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643473 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zz47\" (UniqueName: \"kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.643561 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jzgb\" (UniqueName: \"kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.644652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.645393 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.665400 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zz47\" (UniqueName: \"kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47\") pod \"nova-api-da12-account-create-update-wttp8\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.667166 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x6n6\" (UniqueName: \"kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6\") pod \"nova-cell1-db-create-b624l\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.742719 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.745404 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jzgb\" (UniqueName: \"kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.745550 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.746382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.754649 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.760017 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2010-account-create-update-lfxbp"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.761597 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.767696 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jzgb\" (UniqueName: \"kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb\") pod \"nova-cell0-896e-account-create-update-x9tvp\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.768507 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.770149 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2010-account-create-update-lfxbp"] Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.847896 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.847969 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5vql\" (UniqueName: \"kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.931884 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.949899 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.949963 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5vql\" (UniqueName: \"kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.951004 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:14 crc kubenswrapper[4867]: I1212 07:09:14.970890 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5vql\" (UniqueName: \"kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql\") pod \"nova-cell1-2010-account-create-update-lfxbp\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:15 crc kubenswrapper[4867]: I1212 07:09:15.150884 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:15 crc kubenswrapper[4867]: I1212 07:09:15.882509 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:15 crc kubenswrapper[4867]: I1212 07:09:15.883168 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-log" containerID="cri-o://dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef" gracePeriod=30 Dec 12 07:09:15 crc kubenswrapper[4867]: I1212 07:09:15.883286 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-httpd" containerID="cri-o://6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee" gracePeriod=30 Dec 12 07:09:16 crc kubenswrapper[4867]: I1212 07:09:16.520240 4867 generic.go:334] "Generic (PLEG): container finished" podID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerID="dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef" exitCode=143 Dec 12 07:09:16 crc kubenswrapper[4867]: I1212 07:09:16.520286 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerDied","Data":"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef"} Dec 12 07:09:16 crc kubenswrapper[4867]: I1212 07:09:16.617889 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:16 crc kubenswrapper[4867]: I1212 07:09:16.625738 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-httpd" containerID="cri-o://27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45" gracePeriod=30 Dec 12 07:09:16 crc kubenswrapper[4867]: I1212 07:09:16.625876 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-log" containerID="cri-o://a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6" gracePeriod=30 Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.381802 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.510613 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.510926 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.510966 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.511075 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqr74\" (UniqueName: \"kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.511142 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.511177 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.511200 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts\") pod \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\" (UID: \"d7d510a6-bd18-4ac5-a3b9-1686797ff916\") " Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.512283 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.515183 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.527135 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74" (OuterVolumeSpecName: "kube-api-access-lqr74") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "kube-api-access-lqr74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.529935 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts" (OuterVolumeSpecName: "scripts") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.546096 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7d510a6-bd18-4ac5-a3b9-1686797ff916","Type":"ContainerDied","Data":"a2c54ce60703c602d9d0a200e38c2d9ab3e3f3bae26846b054c60e018e1c1b4e"} Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.546155 4867 scope.go:117] "RemoveContainer" containerID="ef3d627b3369c94b9be6839b951980966127172afc75b31b3abfa1f3a3fb0fed" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.546343 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.570315 4867 generic.go:334] "Generic (PLEG): container finished" podID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerID="a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6" exitCode=143 Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.570359 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerDied","Data":"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6"} Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.596266 4867 scope.go:117] "RemoveContainer" containerID="4c1fdd98d9ae557027eb96d541bae60198656d3677292068ba260ca74e4959e1" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.613429 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.613471 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7d510a6-bd18-4ac5-a3b9-1686797ff916-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.613484 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.613495 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqr74\" (UniqueName: \"kubernetes.io/projected/d7d510a6-bd18-4ac5-a3b9-1686797ff916-kube-api-access-lqr74\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.636570 4867 scope.go:117] "RemoveContainer" containerID="518289223b7438ceaba6563b966ae4a359e3859806f0fefbccb32ce3766718f2" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.654705 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.692460 4867 scope.go:117] "RemoveContainer" containerID="f0e95653ca928d0a0d0624cd8eace10ce6e6530062f0230018c62dd0396df2b4" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.704932 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.716300 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.716341 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.735746 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data" (OuterVolumeSpecName: "config-data") pod "d7d510a6-bd18-4ac5-a3b9-1686797ff916" (UID: "d7d510a6-bd18-4ac5-a3b9-1686797ff916"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.813038 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2010-account-create-update-lfxbp"] Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.817447 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7d510a6-bd18-4ac5-a3b9-1686797ff916-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.940627 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.955994 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.971694 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:17 crc kubenswrapper[4867]: E1212 07:09:17.972219 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="proxy-httpd" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972335 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="proxy-httpd" Dec 12 07:09:17 crc kubenswrapper[4867]: E1212 07:09:17.972350 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="sg-core" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972360 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="sg-core" Dec 12 07:09:17 crc kubenswrapper[4867]: E1212 07:09:17.972394 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-notification-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972404 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-notification-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: E1212 07:09:17.972417 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-central-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972426 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-central-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972662 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="sg-core" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972688 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="proxy-httpd" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972705 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-central-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.972716 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" containerName="ceilometer-notification-agent" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.974752 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.978627 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.978880 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:09:17 crc kubenswrapper[4867]: I1212 07:09:17.997824 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-rchnq"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.022266 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.022658 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.023010 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.023093 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.023507 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.023716 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kctck\" (UniqueName: \"kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.023826 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.027371 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.048311 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tc5wq"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125734 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kctck\" (UniqueName: \"kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125818 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125869 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125895 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.125992 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.126022 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.127054 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.127436 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.131596 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.132871 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.133855 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.137498 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.148818 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kctck\" (UniqueName: \"kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck\") pod \"ceilometer-0\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.211408 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.212288 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.379339 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b624l"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.389079 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.407952 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.416837 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-896e-account-create-update-x9tvp"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.425086 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-da12-account-create-update-wttp8"] Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.602628 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tc5wq" event={"ID":"c6e178c2-60a2-40b1-864d-86fafb2b4d57","Type":"ContainerStarted","Data":"c0e79ee7ccaae63a159eb00b579248378dfebd6adfd1d97ec204a7a06564e0d0"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.607094 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" event={"ID":"dc646b27-e211-4705-91d4-e97fe112855a","Type":"ContainerStarted","Data":"41e04eb1691dc0352e63a7389d8ee835250c0293f3432dd3d95ea9a85070a5f9"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.610494 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-da12-account-create-update-wttp8" event={"ID":"2cf99705-8647-4433-bb42-6869440a63c6","Type":"ContainerStarted","Data":"332ea7ff1f246881411d789b84ef18ad81aea4bcbf0e0aa841d6346ade6d7872"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.612995 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerStarted","Data":"becc0616fc52a7219276f1a0c33ac193b68cf392b6995c852033103fcc75c52e"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.644280 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" event={"ID":"6fced4c1-5bda-4c55-bcf9-80c622893f7d","Type":"ContainerStarted","Data":"eceabc78010dd1dc0c66c2022c93f83ea42d9dae62b652090f320e16bca424b5"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.654522 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"479f0ca7-1408-41d4-abaf-6034d0802100","Type":"ContainerStarted","Data":"48b3ddfeb9f53136e00be49e14a43f05d1fd34eb01693af6aa1fd779aa73a984"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.661890 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rchnq" event={"ID":"0c9cd48c-04e0-4b85-9809-e3c89fbb5081","Type":"ContainerStarted","Data":"551325ab62a11a072efa5bd2478ea10ec51506f583230cbc6747f1d1aeb00cbd"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.681556 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b624l" event={"ID":"8a5b5420-43c8-4f14-b059-42eb6c7b312f","Type":"ContainerStarted","Data":"b7f940d61ba9f8173cd60f1eb399bc61f4d7dbd09b6a3ea6115368e1c326b0b7"} Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.684373 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.793247854 podStartE2EDuration="15.684312114s" podCreationTimestamp="2025-12-12 07:09:03 +0000 UTC" firstStartedPulling="2025-12-12 07:09:04.353107187 +0000 UTC m=+1231.924488446" lastFinishedPulling="2025-12-12 07:09:17.244171437 +0000 UTC m=+1244.815552706" observedRunningTime="2025-12-12 07:09:18.679969427 +0000 UTC m=+1246.251350696" watchObservedRunningTime="2025-12-12 07:09:18.684312114 +0000 UTC m=+1246.255693383" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.712834 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-rchnq" podStartSLOduration=4.712813713 podStartE2EDuration="4.712813713s" podCreationTimestamp="2025-12-12 07:09:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:18.705528625 +0000 UTC m=+1246.276909894" watchObservedRunningTime="2025-12-12 07:09:18.712813713 +0000 UTC m=+1246.284194982" Dec 12 07:09:18 crc kubenswrapper[4867]: I1212 07:09:18.917944 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d510a6-bd18-4ac5-a3b9-1686797ff916" path="/var/lib/kubelet/pods/d7d510a6-bd18-4ac5-a3b9-1686797ff916/volumes" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.131128 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.658439 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.733013 4867 generic.go:334] "Generic (PLEG): container finished" podID="0c9cd48c-04e0-4b85-9809-e3c89fbb5081" containerID="8cc586c884d2d16c09c77de72811d41b3e24fe24ed8d99ac084d0cfb59b68ee4" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.733415 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rchnq" event={"ID":"0c9cd48c-04e0-4b85-9809-e3c89fbb5081","Type":"ContainerDied","Data":"8cc586c884d2d16c09c77de72811d41b3e24fe24ed8d99ac084d0cfb59b68ee4"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.758218 4867 generic.go:334] "Generic (PLEG): container finished" podID="c6e178c2-60a2-40b1-864d-86fafb2b4d57" containerID="f052721f144c8d81984067d872d7c79d7f1cda0b9aede5c27429be698c0098d5" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.758341 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tc5wq" event={"ID":"c6e178c2-60a2-40b1-864d-86fafb2b4d57","Type":"ContainerDied","Data":"f052721f144c8d81984067d872d7c79d7f1cda0b9aede5c27429be698c0098d5"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.767552 4867 generic.go:334] "Generic (PLEG): container finished" podID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerID="6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.767614 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerDied","Data":"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.767646 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72","Type":"ContainerDied","Data":"adf18422e18f264a1ef24f3681bcf076a5e9b573c0330fe953766a004768d5be"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.767667 4867 scope.go:117] "RemoveContainer" containerID="6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.767813 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.772058 4867 generic.go:334] "Generic (PLEG): container finished" podID="dc646b27-e211-4705-91d4-e97fe112855a" containerID="fbc9e2de34a9d4fa8a9c7a372b1fa0f7bd6ae42ab2cbe3e5311d425d777dafd7" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.772135 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" event={"ID":"dc646b27-e211-4705-91d4-e97fe112855a","Type":"ContainerDied","Data":"fbc9e2de34a9d4fa8a9c7a372b1fa0f7bd6ae42ab2cbe3e5311d425d777dafd7"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777075 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777202 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777299 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777392 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777436 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777549 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.777579 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtdx8\" (UniqueName: \"kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8\") pod \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\" (UID: \"f77f99b8-46c3-4b6d-abe7-f6ce348f4d72\") " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.778938 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.782583 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.782965 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs" (OuterVolumeSpecName: "logs") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.817003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerStarted","Data":"8cd180ed7cb6b465bdfd0105a35071f5be7a9810e30b308406669748fe930aab"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.819346 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts" (OuterVolumeSpecName: "scripts") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.829531 4867 generic.go:334] "Generic (PLEG): container finished" podID="6fced4c1-5bda-4c55-bcf9-80c622893f7d" containerID="c73943457de5fe6564da42074030b9d5a1e4173779760522de4f412b52e3c63e" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.829625 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" event={"ID":"6fced4c1-5bda-4c55-bcf9-80c622893f7d","Type":"ContainerDied","Data":"c73943457de5fe6564da42074030b9d5a1e4173779760522de4f412b52e3c63e"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.831028 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8" (OuterVolumeSpecName: "kube-api-access-gtdx8") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "kube-api-access-gtdx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.842215 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a5b5420-43c8-4f14-b059-42eb6c7b312f" containerID="97739a03e3bc94aca021895c2059a73f840207d78f95cbff068baa5aa4b4baf0" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.842461 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b624l" event={"ID":"8a5b5420-43c8-4f14-b059-42eb6c7b312f","Type":"ContainerDied","Data":"97739a03e3bc94aca021895c2059a73f840207d78f95cbff068baa5aa4b4baf0"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.844778 4867 generic.go:334] "Generic (PLEG): container finished" podID="2cf99705-8647-4433-bb42-6869440a63c6" containerID="fd75e02b1ef268a12725da9e2896abd47d835148c1ca4dace74da4dd77e5455d" exitCode=0 Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.844847 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-da12-account-create-update-wttp8" event={"ID":"2cf99705-8647-4433-bb42-6869440a63c6","Type":"ContainerDied","Data":"fd75e02b1ef268a12725da9e2896abd47d835148c1ca4dace74da4dd77e5455d"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.856980 4867 scope.go:117] "RemoveContainer" containerID="dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.862367 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerStarted","Data":"009efa7ff151d899c9c12d5ccad1f687f7819798314685239298b28b18116147"} Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.870144 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880028 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880091 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtdx8\" (UniqueName: \"kubernetes.io/projected/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-kube-api-access-gtdx8\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880119 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880135 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880146 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.880157 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.897625 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.907779 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data" (OuterVolumeSpecName: "config-data") pod "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" (UID: "f77f99b8-46c3-4b6d-abe7-f6ce348f4d72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.912269 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.931073 4867 scope.go:117] "RemoveContainer" containerID="6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.931615 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:09:19 crc kubenswrapper[4867]: E1212 07:09:19.933925 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee\": container with ID starting with 6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee not found: ID does not exist" containerID="6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.933979 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee"} err="failed to get container status \"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee\": rpc error: code = NotFound desc = could not find container \"6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee\": container with ID starting with 6ee41b8ab5f9a733bbc8d9fef8ba7d553a7a032bac271a3a03e235220a7ffcee not found: ID does not exist" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.934004 4867 scope.go:117] "RemoveContainer" containerID="dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef" Dec 12 07:09:19 crc kubenswrapper[4867]: E1212 07:09:19.937382 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef\": container with ID starting with dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef not found: ID does not exist" containerID="dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.937454 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef"} err="failed to get container status \"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef\": rpc error: code = NotFound desc = could not find container \"dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef\": container with ID starting with dc97431b93a257240607dfdbe64bfa1ab8f547f212889d8fb53ff3717c582aef not found: ID does not exist" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.987322 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.987361 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:19 crc kubenswrapper[4867]: I1212 07:09:19.987378 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.047185 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.128144 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.172078 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.210691 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:20 crc kubenswrapper[4867]: E1212 07:09:20.213124 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-httpd" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.213166 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-httpd" Dec 12 07:09:20 crc kubenswrapper[4867]: E1212 07:09:20.213209 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-log" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.213219 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-log" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.218096 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-httpd" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.218149 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" containerName="glance-log" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.224447 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.227389 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.227711 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.259720 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327506 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327598 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327641 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wsd7\" (UniqueName: \"kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327707 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327734 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327790 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327865 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.327893 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.429816 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430175 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430243 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430316 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430337 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430404 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430443 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.430468 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wsd7\" (UniqueName: \"kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.431361 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.440157 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.446202 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.447149 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.447265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.465497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.472558 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.477007 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wsd7\" (UniqueName: \"kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.524287 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.589685 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.598823 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.633839 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gphg6\" (UniqueName: \"kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634351 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634381 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634406 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634443 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634474 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634491 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.634586 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle\") pod \"3844a91f-5305-40f8-80a8-7801706fe0b4\" (UID: \"3844a91f-5305-40f8-80a8-7801706fe0b4\") " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.639485 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs" (OuterVolumeSpecName: "logs") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.639585 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.641735 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts" (OuterVolumeSpecName: "scripts") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.641995 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6" (OuterVolumeSpecName: "kube-api-access-gphg6") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "kube-api-access-gphg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.646947 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.685464 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.736492 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data" (OuterVolumeSpecName: "config-data") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737179 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737205 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gphg6\" (UniqueName: \"kubernetes.io/projected/3844a91f-5305-40f8-80a8-7801706fe0b4-kube-api-access-gphg6\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737219 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737262 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737272 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737282 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3844a91f-5305-40f8-80a8-7801706fe0b4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.737352 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.787966 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.790902 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3844a91f-5305-40f8-80a8-7801706fe0b4" (UID: "3844a91f-5305-40f8-80a8-7801706fe0b4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.841942 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3844a91f-5305-40f8-80a8-7801706fe0b4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.842319 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.863419 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f77f99b8-46c3-4b6d-abe7-f6ce348f4d72" path="/var/lib/kubelet/pods/f77f99b8-46c3-4b6d-abe7-f6ce348f4d72/volumes" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.894320 4867 generic.go:334] "Generic (PLEG): container finished" podID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerID="27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45" exitCode=0 Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.894430 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.895208 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerDied","Data":"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45"} Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.895279 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3844a91f-5305-40f8-80a8-7801706fe0b4","Type":"ContainerDied","Data":"52314552ac36652bea8ffdce9e8c5ddb8e12c6d6a7bd8adb8811dc4750fc41d2"} Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.895302 4867 scope.go:117] "RemoveContainer" containerID="27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.921149 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerStarted","Data":"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf"} Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.967780 4867 scope.go:117] "RemoveContainer" containerID="a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6" Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.981316 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:20 crc kubenswrapper[4867]: I1212 07:09:20.993316 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.004170 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:21 crc kubenswrapper[4867]: E1212 07:09:21.004606 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-log" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.004622 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-log" Dec 12 07:09:21 crc kubenswrapper[4867]: E1212 07:09:21.004669 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-httpd" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.004677 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-httpd" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.004871 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-log" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.004906 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" containerName="glance-httpd" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.006967 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.014613 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.015460 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.017759 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.080945 4867 scope.go:117] "RemoveContainer" containerID="27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45" Dec 12 07:09:21 crc kubenswrapper[4867]: E1212 07:09:21.081484 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45\": container with ID starting with 27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45 not found: ID does not exist" containerID="27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.081516 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45"} err="failed to get container status \"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45\": rpc error: code = NotFound desc = could not find container \"27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45\": container with ID starting with 27a9a97f64d4fa0c20780d1481d4681c0ed06b3c9ece9f4e05c7f3bcc0db8c45 not found: ID does not exist" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.081542 4867 scope.go:117] "RemoveContainer" containerID="a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6" Dec 12 07:09:21 crc kubenswrapper[4867]: E1212 07:09:21.081884 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6\": container with ID starting with a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6 not found: ID does not exist" containerID="a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.081921 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6"} err="failed to get container status \"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6\": rpc error: code = NotFound desc = could not find container \"a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6\": container with ID starting with a115e1ab1095f7f08b39171456a532bbb91dce91e910e0697a8b42c57fa136a6 not found: ID does not exist" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.184898 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185262 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztx68\" (UniqueName: \"kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185462 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185591 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185709 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185806 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.185904 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.186035 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztx68\" (UniqueName: \"kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287530 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287583 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287610 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287637 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287659 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287708 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.287739 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.289042 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.289129 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.291967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.298146 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.316179 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.317886 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.320967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.322693 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztx68\" (UniqueName: \"kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.331207 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.370819 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.402751 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.561217 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.701385 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts\") pod \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.701511 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb4nj\" (UniqueName: \"kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj\") pod \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\" (UID: \"c6e178c2-60a2-40b1-864d-86fafb2b4d57\") " Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.702703 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c6e178c2-60a2-40b1-864d-86fafb2b4d57" (UID: "c6e178c2-60a2-40b1-864d-86fafb2b4d57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.737679 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj" (OuterVolumeSpecName: "kube-api-access-nb4nj") pod "c6e178c2-60a2-40b1-864d-86fafb2b4d57" (UID: "c6e178c2-60a2-40b1-864d-86fafb2b4d57"). InnerVolumeSpecName "kube-api-access-nb4nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.803619 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c6e178c2-60a2-40b1-864d-86fafb2b4d57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.803661 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb4nj\" (UniqueName: \"kubernetes.io/projected/c6e178c2-60a2-40b1-864d-86fafb2b4d57-kube-api-access-nb4nj\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:21 crc kubenswrapper[4867]: I1212 07:09:21.955938 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:21.996522 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.023797 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts\") pod \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.024271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cswv5\" (UniqueName: \"kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5\") pod \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\" (UID: \"0c9cd48c-04e0-4b85-9809-e3c89fbb5081\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.032360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c9cd48c-04e0-4b85-9809-e3c89fbb5081" (UID: "0c9cd48c-04e0-4b85-9809-e3c89fbb5081"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.080511 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5" (OuterVolumeSpecName: "kube-api-access-cswv5") pod "0c9cd48c-04e0-4b85-9809-e3c89fbb5081" (UID: "0c9cd48c-04e0-4b85-9809-e3c89fbb5081"). InnerVolumeSpecName "kube-api-access-cswv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.129711 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5vql\" (UniqueName: \"kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql\") pod \"dc646b27-e211-4705-91d4-e97fe112855a\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.130000 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts\") pod \"dc646b27-e211-4705-91d4-e97fe112855a\" (UID: \"dc646b27-e211-4705-91d4-e97fe112855a\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.130553 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cswv5\" (UniqueName: \"kubernetes.io/projected/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-kube-api-access-cswv5\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.130579 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c9cd48c-04e0-4b85-9809-e3c89fbb5081-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.136502 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql" (OuterVolumeSpecName: "kube-api-access-b5vql") pod "dc646b27-e211-4705-91d4-e97fe112855a" (UID: "dc646b27-e211-4705-91d4-e97fe112855a"). InnerVolumeSpecName "kube-api-access-b5vql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.142512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc646b27-e211-4705-91d4-e97fe112855a" (UID: "dc646b27-e211-4705-91d4-e97fe112855a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.152841 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerStarted","Data":"688b563243ecf18f051dc395d4e9bab540550c17791a5cf14397013544eebe33"} Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.155422 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.157518 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-rchnq" event={"ID":"0c9cd48c-04e0-4b85-9809-e3c89fbb5081","Type":"ContainerDied","Data":"551325ab62a11a072efa5bd2478ea10ec51506f583230cbc6747f1d1aeb00cbd"} Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.157561 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="551325ab62a11a072efa5bd2478ea10ec51506f583230cbc6747f1d1aeb00cbd" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.157629 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-rchnq" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.198218 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tc5wq" event={"ID":"c6e178c2-60a2-40b1-864d-86fafb2b4d57","Type":"ContainerDied","Data":"c0e79ee7ccaae63a159eb00b579248378dfebd6adfd1d97ec204a7a06564e0d0"} Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.198286 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0e79ee7ccaae63a159eb00b579248378dfebd6adfd1d97ec204a7a06564e0d0" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.198437 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.199310 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tc5wq" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.201018 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.209270 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" event={"ID":"dc646b27-e211-4705-91d4-e97fe112855a","Type":"ContainerDied","Data":"41e04eb1691dc0352e63a7389d8ee835250c0293f3432dd3d95ea9a85070a5f9"} Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.209338 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41e04eb1691dc0352e63a7389d8ee835250c0293f3432dd3d95ea9a85070a5f9" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.212393 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2010-account-create-update-lfxbp" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.235642 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jzgb\" (UniqueName: \"kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb\") pod \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.235838 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts\") pod \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.236125 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts\") pod \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\" (UID: \"6fced4c1-5bda-4c55-bcf9-80c622893f7d\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.236212 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts\") pod \"2cf99705-8647-4433-bb42-6869440a63c6\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.236409 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x6n6\" (UniqueName: \"kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6\") pod \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\" (UID: \"8a5b5420-43c8-4f14-b059-42eb6c7b312f\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.236464 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zz47\" (UniqueName: \"kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47\") pod \"2cf99705-8647-4433-bb42-6869440a63c6\" (UID: \"2cf99705-8647-4433-bb42-6869440a63c6\") " Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.237035 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5vql\" (UniqueName: \"kubernetes.io/projected/dc646b27-e211-4705-91d4-e97fe112855a-kube-api-access-b5vql\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.237053 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc646b27-e211-4705-91d4-e97fe112855a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.237638 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2cf99705-8647-4433-bb42-6869440a63c6" (UID: "2cf99705-8647-4433-bb42-6869440a63c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.237821 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fced4c1-5bda-4c55-bcf9-80c622893f7d" (UID: "6fced4c1-5bda-4c55-bcf9-80c622893f7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.249033 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerStarted","Data":"82e68b2b9f69ed30e1eee05edc30a399c219b479de169d6d7373d0b615332b11"} Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.253553 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a5b5420-43c8-4f14-b059-42eb6c7b312f" (UID: "8a5b5420-43c8-4f14-b059-42eb6c7b312f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.271561 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6" (OuterVolumeSpecName: "kube-api-access-5x6n6") pod "8a5b5420-43c8-4f14-b059-42eb6c7b312f" (UID: "8a5b5420-43c8-4f14-b059-42eb6c7b312f"). InnerVolumeSpecName "kube-api-access-5x6n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.278287 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb" (OuterVolumeSpecName: "kube-api-access-8jzgb") pod "6fced4c1-5bda-4c55-bcf9-80c622893f7d" (UID: "6fced4c1-5bda-4c55-bcf9-80c622893f7d"). InnerVolumeSpecName "kube-api-access-8jzgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.279509 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47" (OuterVolumeSpecName: "kube-api-access-5zz47") pod "2cf99705-8647-4433-bb42-6869440a63c6" (UID: "2cf99705-8647-4433-bb42-6869440a63c6"). InnerVolumeSpecName "kube-api-access-5zz47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344158 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jzgb\" (UniqueName: \"kubernetes.io/projected/6fced4c1-5bda-4c55-bcf9-80c622893f7d-kube-api-access-8jzgb\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344186 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a5b5420-43c8-4f14-b059-42eb6c7b312f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344198 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fced4c1-5bda-4c55-bcf9-80c622893f7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344209 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cf99705-8647-4433-bb42-6869440a63c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344222 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x6n6\" (UniqueName: \"kubernetes.io/projected/8a5b5420-43c8-4f14-b059-42eb6c7b312f-kube-api-access-5x6n6\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.344249 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zz47\" (UniqueName: \"kubernetes.io/projected/2cf99705-8647-4433-bb42-6869440a63c6-kube-api-access-5zz47\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.369617 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.374964 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=12.374939433 podStartE2EDuration="12.374939433s" podCreationTimestamp="2025-12-12 07:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:22.30157666 +0000 UTC m=+1249.872957949" watchObservedRunningTime="2025-12-12 07:09:22.374939433 +0000 UTC m=+1249.946320702" Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.591861 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:09:22 crc kubenswrapper[4867]: W1212 07:09:22.602126 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50a1fd31_a115_4bb6_b65e_70adc5dd77ed.slice/crio-753cac05b80813598a54d00c3be296ae0bf594560258313891cd470db8a16304 WatchSource:0}: Error finding container 753cac05b80813598a54d00c3be296ae0bf594560258313891cd470db8a16304: Status 404 returned error can't find the container with id 753cac05b80813598a54d00c3be296ae0bf594560258313891cd470db8a16304 Dec 12 07:09:22 crc kubenswrapper[4867]: I1212 07:09:22.867303 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3844a91f-5305-40f8-80a8-7801706fe0b4" path="/var/lib/kubelet/pods/3844a91f-5305-40f8-80a8-7801706fe0b4/volumes" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.277534 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerStarted","Data":"753cac05b80813598a54d00c3be296ae0bf594560258313891cd470db8a16304"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.283231 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b624l" event={"ID":"8a5b5420-43c8-4f14-b059-42eb6c7b312f","Type":"ContainerDied","Data":"b7f940d61ba9f8173cd60f1eb399bc61f4d7dbd09b6a3ea6115368e1c326b0b7"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.283288 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7f940d61ba9f8173cd60f1eb399bc61f4d7dbd09b6a3ea6115368e1c326b0b7" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.283350 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b624l" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.290729 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-da12-account-create-update-wttp8" event={"ID":"2cf99705-8647-4433-bb42-6869440a63c6","Type":"ContainerDied","Data":"332ea7ff1f246881411d789b84ef18ad81aea4bcbf0e0aa841d6346ade6d7872"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.290793 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="332ea7ff1f246881411d789b84ef18ad81aea4bcbf0e0aa841d6346ade6d7872" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.290889 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-da12-account-create-update-wttp8" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.312245 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerStarted","Data":"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.312298 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerStarted","Data":"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.316231 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" event={"ID":"6fced4c1-5bda-4c55-bcf9-80c622893f7d","Type":"ContainerDied","Data":"eceabc78010dd1dc0c66c2022c93f83ea42d9dae62b652090f320e16bca424b5"} Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.316287 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eceabc78010dd1dc0c66c2022c93f83ea42d9dae62b652090f320e16bca424b5" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.316341 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-896e-account-create-update-x9tvp" Dec 12 07:09:23 crc kubenswrapper[4867]: I1212 07:09:23.319682 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerStarted","Data":"e0e23c15cb95fadeb8c5fb6992fb31533b6edb8b32c8f954630c0f55ff2e3c91"} Dec 12 07:09:24 crc kubenswrapper[4867]: I1212 07:09:24.330087 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerStarted","Data":"506b0e61cb33a6fbdd013a5cdbf2ecb8a9e984240941a905449f20e6d9bda9a1"} Dec 12 07:09:24 crc kubenswrapper[4867]: I1212 07:09:24.332680 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerStarted","Data":"4a8ad42ecdb710efacb437807e79af57dea08fcccf32bfdc46ca2fca80f02ba9"} Dec 12 07:09:24 crc kubenswrapper[4867]: I1212 07:09:24.363364 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.363341737 podStartE2EDuration="4.363341737s" podCreationTimestamp="2025-12-12 07:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:24.35941821 +0000 UTC m=+1251.930799479" watchObservedRunningTime="2025-12-12 07:09:24.363341737 +0000 UTC m=+1251.934723006" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.052035 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.074254 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xkgvg"] Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.074884 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6e178c2-60a2-40b1-864d-86fafb2b4d57" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.074984 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6e178c2-60a2-40b1-864d-86fafb2b4d57" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.075053 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fced4c1-5bda-4c55-bcf9-80c622893f7d" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075104 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fced4c1-5bda-4c55-bcf9-80c622893f7d" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.075166 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5b5420-43c8-4f14-b059-42eb6c7b312f" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075219 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5b5420-43c8-4f14-b059-42eb6c7b312f" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.075331 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9cd48c-04e0-4b85-9809-e3c89fbb5081" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075395 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9cd48c-04e0-4b85-9809-e3c89fbb5081" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.075459 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc646b27-e211-4705-91d4-e97fe112855a" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075510 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc646b27-e211-4705-91d4-e97fe112855a" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: E1212 07:09:25.075570 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf99705-8647-4433-bb42-6869440a63c6" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075621 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf99705-8647-4433-bb42-6869440a63c6" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075832 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fced4c1-5bda-4c55-bcf9-80c622893f7d" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075901 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf99705-8647-4433-bb42-6869440a63c6" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.075962 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc646b27-e211-4705-91d4-e97fe112855a" containerName="mariadb-account-create-update" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.076016 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5b5420-43c8-4f14-b059-42eb6c7b312f" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.076075 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9cd48c-04e0-4b85-9809-e3c89fbb5081" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.076131 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6e178c2-60a2-40b1-864d-86fafb2b4d57" containerName="mariadb-database-create" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.076784 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.078422 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.078503 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.078746 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l679z" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.102885 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xkgvg"] Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.106236 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.106436 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6v7v\" (UniqueName: \"kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.106487 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.106578 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.207481 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.207565 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.207649 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6v7v\" (UniqueName: \"kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.207689 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.220223 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.224877 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.260104 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.272077 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6v7v\" (UniqueName: \"kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v\") pod \"nova-cell0-conductor-db-sync-xkgvg\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.399045 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerStarted","Data":"f164d088b657a841b64a09bed96dee589e4f7130244d17533d2c9a7d6fdb6dab"} Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.399222 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.513560 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.513539791 podStartE2EDuration="5.513539791s" podCreationTimestamp="2025-12-12 07:09:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:25.478785367 +0000 UTC m=+1253.050166646" watchObservedRunningTime="2025-12-12 07:09:25.513539791 +0000 UTC m=+1253.084921060" Dec 12 07:09:25 crc kubenswrapper[4867]: I1212 07:09:25.907762 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.105736 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xkgvg"] Dec 12 07:09:26 crc kubenswrapper[4867]: W1212 07:09:26.106931 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod714a3a25_e3f2_4847_9200_0c633bc0254f.slice/crio-dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021 WatchSource:0}: Error finding container dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021: Status 404 returned error can't find the container with id dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021 Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.180323 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411637 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerStarted","Data":"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54"} Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411771 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-central-agent" containerID="cri-o://fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf" gracePeriod=30 Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411797 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="proxy-httpd" containerID="cri-o://d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54" gracePeriod=30 Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411807 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="sg-core" containerID="cri-o://fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830" gracePeriod=30 Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411810 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.411920 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-notification-agent" containerID="cri-o://a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5" gracePeriod=30 Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.414162 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" event={"ID":"714a3a25-e3f2-4847-9200-0c633bc0254f","Type":"ContainerStarted","Data":"dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021"} Dec 12 07:09:26 crc kubenswrapper[4867]: I1212 07:09:26.445230 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.325281443 podStartE2EDuration="9.445206207s" podCreationTimestamp="2025-12-12 07:09:17 +0000 UTC" firstStartedPulling="2025-12-12 07:09:19.235733008 +0000 UTC m=+1246.807114277" lastFinishedPulling="2025-12-12 07:09:25.355657782 +0000 UTC m=+1252.927039041" observedRunningTime="2025-12-12 07:09:26.43841833 +0000 UTC m=+1254.009799599" watchObservedRunningTime="2025-12-12 07:09:26.445206207 +0000 UTC m=+1254.016587476" Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.429615 4867 generic.go:334] "Generic (PLEG): container finished" podID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerID="d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54" exitCode=0 Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.429990 4867 generic.go:334] "Generic (PLEG): container finished" podID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerID="fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830" exitCode=2 Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.430004 4867 generic.go:334] "Generic (PLEG): container finished" podID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerID="a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5" exitCode=0 Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.429695 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerDied","Data":"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54"} Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.430048 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerDied","Data":"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830"} Dec 12 07:09:27 crc kubenswrapper[4867]: I1212 07:09:27.430069 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerDied","Data":"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5"} Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.167205 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.300724 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.300945 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301137 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kctck\" (UniqueName: \"kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301180 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301248 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301667 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301896 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.301943 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle\") pod \"f0b70548-2efa-4444-ab2c-c9d326f3959d\" (UID: \"f0b70548-2efa-4444-ab2c-c9d326f3959d\") " Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.302578 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.302588 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.317164 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck" (OuterVolumeSpecName: "kube-api-access-kctck") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "kube-api-access-kctck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.324433 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts" (OuterVolumeSpecName: "scripts") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.403972 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kctck\" (UniqueName: \"kubernetes.io/projected/f0b70548-2efa-4444-ab2c-c9d326f3959d-kube-api-access-kctck\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.404013 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0b70548-2efa-4444-ab2c-c9d326f3959d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.404025 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.413373 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.426041 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.465352 4867 generic.go:334] "Generic (PLEG): container finished" podID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerID="fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf" exitCode=0 Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.465404 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerDied","Data":"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf"} Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.465428 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.465448 4867 scope.go:117] "RemoveContainer" containerID="d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.465435 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0b70548-2efa-4444-ab2c-c9d326f3959d","Type":"ContainerDied","Data":"009efa7ff151d899c9c12d5ccad1f687f7819798314685239298b28b18116147"} Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.471923 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data" (OuterVolumeSpecName: "config-data") pod "f0b70548-2efa-4444-ab2c-c9d326f3959d" (UID: "f0b70548-2efa-4444-ab2c-c9d326f3959d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.493006 4867 scope.go:117] "RemoveContainer" containerID="fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.506132 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.506175 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.506188 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b70548-2efa-4444-ab2c-c9d326f3959d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.516131 4867 scope.go:117] "RemoveContainer" containerID="a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.540036 4867 scope.go:117] "RemoveContainer" containerID="fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.560645 4867 scope.go:117] "RemoveContainer" containerID="d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.561135 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54\": container with ID starting with d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54 not found: ID does not exist" containerID="d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.561180 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54"} err="failed to get container status \"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54\": rpc error: code = NotFound desc = could not find container \"d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54\": container with ID starting with d569d2c0c1a598f1e4b28c1ad941f3e4d8a399904b6ca059f0217c6f1051ca54 not found: ID does not exist" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.561206 4867 scope.go:117] "RemoveContainer" containerID="fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.561609 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830\": container with ID starting with fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830 not found: ID does not exist" containerID="fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.561646 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830"} err="failed to get container status \"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830\": rpc error: code = NotFound desc = could not find container \"fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830\": container with ID starting with fcda2eb6a21f500089049b3889966ee3d5e10e65fa535049429837accf69e830 not found: ID does not exist" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.561670 4867 scope.go:117] "RemoveContainer" containerID="a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.561987 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5\": container with ID starting with a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5 not found: ID does not exist" containerID="a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.562034 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5"} err="failed to get container status \"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5\": rpc error: code = NotFound desc = could not find container \"a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5\": container with ID starting with a045510bbb922a177242a82c47dc800aadd97bcf8ff7fb7e324d895ef71f1bc5 not found: ID does not exist" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.562059 4867 scope.go:117] "RemoveContainer" containerID="fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.562501 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf\": container with ID starting with fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf not found: ID does not exist" containerID="fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.562565 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf"} err="failed to get container status \"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf\": rpc error: code = NotFound desc = could not find container \"fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf\": container with ID starting with fed96004398ac304c53f89a441383954a207c8044a160adc73346a8353002bcf not found: ID does not exist" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.858554 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.874869 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.887708 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.888119 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="sg-core" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888133 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="sg-core" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.888157 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-central-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888172 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-central-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.888182 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-notification-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888188 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-notification-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: E1212 07:09:28.888221 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="proxy-httpd" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888232 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="proxy-httpd" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888420 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-central-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888438 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="ceilometer-notification-agent" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888449 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="proxy-httpd" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.888459 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" containerName="sg-core" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.890308 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.893171 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.895461 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.896829 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.988777 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.989172 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.989222 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.990072 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:09:28 crc kubenswrapper[4867]: I1212 07:09:28.990136 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97" gracePeriod=600 Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.014929 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.014973 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fpfj\" (UniqueName: \"kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.015030 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.015048 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.015076 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.015279 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.015527 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.117682 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.117850 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.117917 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.117949 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fpfj\" (UniqueName: \"kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.118014 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.118036 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.118070 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.118575 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.120697 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.122155 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.122479 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.125074 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.133429 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.140000 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fpfj\" (UniqueName: \"kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj\") pod \"ceilometer-0\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.214889 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.484723 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97" exitCode=0 Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.485168 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97"} Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.485196 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d"} Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.485213 4867 scope.go:117] "RemoveContainer" containerID="b62cd3b8f56e87ac9166bd99a70e6658b6bf34dcb606642c25ce818ce7d82713" Dec 12 07:09:29 crc kubenswrapper[4867]: I1212 07:09:29.738152 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:30 crc kubenswrapper[4867]: I1212 07:09:30.600177 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:30 crc kubenswrapper[4867]: I1212 07:09:30.600539 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:30 crc kubenswrapper[4867]: I1212 07:09:30.635883 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:30 crc kubenswrapper[4867]: I1212 07:09:30.649986 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:30 crc kubenswrapper[4867]: I1212 07:09:30.871005 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b70548-2efa-4444-ab2c-c9d326f3959d" path="/var/lib/kubelet/pods/f0b70548-2efa-4444-ab2c-c9d326f3959d/volumes" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.403767 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.404110 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.443642 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.451687 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.526362 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.526411 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.526423 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.526433 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.826076 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.886685 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.888193 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c7bdbbdf4-sbn95" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-api" containerID="cri-o://5ee0aa180aea7ab1ad6875c0ea5b1b61058a02dcd72288d4cf0df2058299c394" gracePeriod=30 Dec 12 07:09:31 crc kubenswrapper[4867]: I1212 07:09:31.888341 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6c7bdbbdf4-sbn95" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-httpd" containerID="cri-o://4f9fc24b43d1f00ce0c09640b3c2ab26e264143888b3092794852d66e2f6fd8b" gracePeriod=30 Dec 12 07:09:32 crc kubenswrapper[4867]: I1212 07:09:32.538772 4867 generic.go:334] "Generic (PLEG): container finished" podID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerID="4f9fc24b43d1f00ce0c09640b3c2ab26e264143888b3092794852d66e2f6fd8b" exitCode=0 Dec 12 07:09:32 crc kubenswrapper[4867]: I1212 07:09:32.538883 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerDied","Data":"4f9fc24b43d1f00ce0c09640b3c2ab26e264143888b3092794852d66e2f6fd8b"} Dec 12 07:09:33 crc kubenswrapper[4867]: I1212 07:09:33.957022 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:33 crc kubenswrapper[4867]: I1212 07:09:33.957490 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:09:34 crc kubenswrapper[4867]: I1212 07:09:34.004920 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 07:09:35 crc kubenswrapper[4867]: I1212 07:09:35.521954 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 07:09:35 crc kubenswrapper[4867]: I1212 07:09:35.522767 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 07:09:35 crc kubenswrapper[4867]: I1212 07:09:35.531208 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 07:09:38 crc kubenswrapper[4867]: I1212 07:09:38.476789 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:38 crc kubenswrapper[4867]: I1212 07:09:38.604721 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerStarted","Data":"7a503687a9318d60a894d5d45ac8ad0f328e7ac445dbce0ff341d430b211b361"} Dec 12 07:09:39 crc kubenswrapper[4867]: I1212 07:09:39.629958 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerStarted","Data":"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400"} Dec 12 07:09:39 crc kubenswrapper[4867]: I1212 07:09:39.639491 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" event={"ID":"714a3a25-e3f2-4847-9200-0c633bc0254f","Type":"ContainerStarted","Data":"2c02009ab1e1505ffea6c57acadcdd4d133ceb5623c0d8e451ec9680ea621aeb"} Dec 12 07:09:39 crc kubenswrapper[4867]: I1212 07:09:39.662591 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" podStartSLOduration=2.255403399 podStartE2EDuration="14.662565277s" podCreationTimestamp="2025-12-12 07:09:25 +0000 UTC" firstStartedPulling="2025-12-12 07:09:26.110104305 +0000 UTC m=+1253.681485574" lastFinishedPulling="2025-12-12 07:09:38.517266173 +0000 UTC m=+1266.088647452" observedRunningTime="2025-12-12 07:09:39.6537584 +0000 UTC m=+1267.225139669" watchObservedRunningTime="2025-12-12 07:09:39.662565277 +0000 UTC m=+1267.233946556" Dec 12 07:09:40 crc kubenswrapper[4867]: I1212 07:09:40.665826 4867 generic.go:334] "Generic (PLEG): container finished" podID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerID="5ee0aa180aea7ab1ad6875c0ea5b1b61058a02dcd72288d4cf0df2058299c394" exitCode=0 Dec 12 07:09:40 crc kubenswrapper[4867]: I1212 07:09:40.666481 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerDied","Data":"5ee0aa180aea7ab1ad6875c0ea5b1b61058a02dcd72288d4cf0df2058299c394"} Dec 12 07:09:40 crc kubenswrapper[4867]: I1212 07:09:40.673876 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerStarted","Data":"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c"} Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.132591 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.200444 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config\") pod \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.200527 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f9wk\" (UniqueName: \"kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk\") pod \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.200791 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config\") pod \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.200834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle\") pod \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.200872 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs\") pod \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\" (UID: \"bd553e42-0591-43e4-8ab3-bbe2ebbb7379\") " Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.208634 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "bd553e42-0591-43e4-8ab3-bbe2ebbb7379" (UID: "bd553e42-0591-43e4-8ab3-bbe2ebbb7379"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.221500 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk" (OuterVolumeSpecName: "kube-api-access-4f9wk") pod "bd553e42-0591-43e4-8ab3-bbe2ebbb7379" (UID: "bd553e42-0591-43e4-8ab3-bbe2ebbb7379"). InnerVolumeSpecName "kube-api-access-4f9wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.290842 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config" (OuterVolumeSpecName: "config") pod "bd553e42-0591-43e4-8ab3-bbe2ebbb7379" (UID: "bd553e42-0591-43e4-8ab3-bbe2ebbb7379"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.302109 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.302163 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f9wk\" (UniqueName: \"kubernetes.io/projected/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-kube-api-access-4f9wk\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.302179 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.303353 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd553e42-0591-43e4-8ab3-bbe2ebbb7379" (UID: "bd553e42-0591-43e4-8ab3-bbe2ebbb7379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.311376 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "bd553e42-0591-43e4-8ab3-bbe2ebbb7379" (UID: "bd553e42-0591-43e4-8ab3-bbe2ebbb7379"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.403434 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.403808 4867 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd553e42-0591-43e4-8ab3-bbe2ebbb7379-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.687026 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c7bdbbdf4-sbn95" event={"ID":"bd553e42-0591-43e4-8ab3-bbe2ebbb7379","Type":"ContainerDied","Data":"c811c84680ea4faba42aafce3b8064b9916234e79bc5afdfd21586aad101aaad"} Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.687104 4867 scope.go:117] "RemoveContainer" containerID="4f9fc24b43d1f00ce0c09640b3c2ab26e264143888b3092794852d66e2f6fd8b" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.687063 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c7bdbbdf4-sbn95" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.694277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerStarted","Data":"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff"} Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.723789 4867 scope.go:117] "RemoveContainer" containerID="5ee0aa180aea7ab1ad6875c0ea5b1b61058a02dcd72288d4cf0df2058299c394" Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.742178 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:09:41 crc kubenswrapper[4867]: I1212 07:09:41.753604 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6c7bdbbdf4-sbn95"] Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.705665 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerStarted","Data":"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13"} Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.706089 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.705886 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="sg-core" containerID="cri-o://911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff" gracePeriod=30 Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.705834 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="proxy-httpd" containerID="cri-o://8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13" gracePeriod=30 Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.705892 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-notification-agent" containerID="cri-o://90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c" gracePeriod=30 Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.705807 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-central-agent" containerID="cri-o://bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400" gracePeriod=30 Dec 12 07:09:42 crc kubenswrapper[4867]: I1212 07:09:42.736895 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.885408696 podStartE2EDuration="14.736875126s" podCreationTimestamp="2025-12-12 07:09:28 +0000 UTC" firstStartedPulling="2025-12-12 07:09:38.407799033 +0000 UTC m=+1265.979180302" lastFinishedPulling="2025-12-12 07:09:42.259265463 +0000 UTC m=+1269.830646732" observedRunningTime="2025-12-12 07:09:42.733708898 +0000 UTC m=+1270.305090167" watchObservedRunningTime="2025-12-12 07:09:42.736875126 +0000 UTC m=+1270.308256395" Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.084586 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" path="/var/lib/kubelet/pods/bd553e42-0591-43e4-8ab3-bbe2ebbb7379/volumes" Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723195 4867 generic.go:334] "Generic (PLEG): container finished" podID="3576089d-8a48-4d20-9a43-acb95e60d037" containerID="8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13" exitCode=0 Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723249 4867 generic.go:334] "Generic (PLEG): container finished" podID="3576089d-8a48-4d20-9a43-acb95e60d037" containerID="911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff" exitCode=2 Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723262 4867 generic.go:334] "Generic (PLEG): container finished" podID="3576089d-8a48-4d20-9a43-acb95e60d037" containerID="90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c" exitCode=0 Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723269 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerDied","Data":"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13"} Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723324 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerDied","Data":"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff"} Dec 12 07:09:43 crc kubenswrapper[4867]: I1212 07:09:43.723342 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerDied","Data":"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c"} Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.053754 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.116116 4867 generic.go:334] "Generic (PLEG): container finished" podID="3576089d-8a48-4d20-9a43-acb95e60d037" containerID="bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400" exitCode=0 Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.116654 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerDied","Data":"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400"} Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.116682 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3576089d-8a48-4d20-9a43-acb95e60d037","Type":"ContainerDied","Data":"7a503687a9318d60a894d5d45ac8ad0f328e7ac445dbce0ff341d430b211b361"} Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.116702 4867 scope.go:117] "RemoveContainer" containerID="8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.116841 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.154756 4867 scope.go:117] "RemoveContainer" containerID="911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.174896 4867 scope.go:117] "RemoveContainer" containerID="90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.199817 4867 scope.go:117] "RemoveContainer" containerID="bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.219472 4867 scope.go:117] "RemoveContainer" containerID="8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220074 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fpfj\" (UniqueName: \"kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220153 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220204 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220249 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220356 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220421 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd\") pod \"3576089d-8a48-4d20-9a43-acb95e60d037\" (UID: \"3576089d-8a48-4d20-9a43-acb95e60d037\") " Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220855 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.220952 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.222636 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13\": container with ID starting with 8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13 not found: ID does not exist" containerID="8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.222667 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13"} err="failed to get container status \"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13\": rpc error: code = NotFound desc = could not find container \"8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13\": container with ID starting with 8f717b739248fe2a2170d43ce0e4bf756137c784602f891f806ab9f3e3bf5e13 not found: ID does not exist" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.222691 4867 scope.go:117] "RemoveContainer" containerID="911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.226466 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff\": container with ID starting with 911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff not found: ID does not exist" containerID="911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.226644 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff"} err="failed to get container status \"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff\": rpc error: code = NotFound desc = could not find container \"911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff\": container with ID starting with 911c3485716ce8f9d952b3460d48d1ebdc78f9f82e63dfcc21833f2c5ae8d9ff not found: ID does not exist" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.226681 4867 scope.go:117] "RemoveContainer" containerID="90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.228399 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c\": container with ID starting with 90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c not found: ID does not exist" containerID="90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.228444 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts" (OuterVolumeSpecName: "scripts") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.228450 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c"} err="failed to get container status \"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c\": rpc error: code = NotFound desc = could not find container \"90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c\": container with ID starting with 90f275e6766d2c09e7694e25eb5ae7c5420b7ee27b5bfeabed9a1d90b4c20b5c not found: ID does not exist" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.228520 4867 scope.go:117] "RemoveContainer" containerID="bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.228839 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400\": container with ID starting with bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400 not found: ID does not exist" containerID="bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.228868 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400"} err="failed to get container status \"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400\": rpc error: code = NotFound desc = could not find container \"bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400\": container with ID starting with bdda8d56f5d052db67c2fd0bf402410836a332a8058ae398012c1cc876392400 not found: ID does not exist" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.247492 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj" (OuterVolumeSpecName: "kube-api-access-2fpfj") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "kube-api-access-2fpfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.256402 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.324405 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.324444 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fpfj\" (UniqueName: \"kubernetes.io/projected/3576089d-8a48-4d20-9a43-acb95e60d037-kube-api-access-2fpfj\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.324456 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.324464 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3576089d-8a48-4d20-9a43-acb95e60d037-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.324473 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.346495 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.360996 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data" (OuterVolumeSpecName: "config-data") pod "3576089d-8a48-4d20-9a43-acb95e60d037" (UID: "3576089d-8a48-4d20-9a43-acb95e60d037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.425725 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.425761 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3576089d-8a48-4d20-9a43-acb95e60d037-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.452695 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.462687 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.487484 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.487911 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-api" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.487928 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-api" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.487950 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="proxy-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.487956 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="proxy-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.487968 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.487974 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.487990 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-central-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.487996 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-central-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.488008 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-notification-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488013 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-notification-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: E1212 07:09:52.488030 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="sg-core" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488037 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="sg-core" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488202 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="sg-core" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488214 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-notification-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488247 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="ceilometer-central-agent" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488260 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488276 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd553e42-0591-43e4-8ab3-bbe2ebbb7379" containerName="neutron-api" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.488289 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" containerName="proxy-httpd" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.490355 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.493624 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.493875 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.511053 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630100 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630178 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630645 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630726 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.630772 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv6zw\" (UniqueName: \"kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732213 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732340 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732380 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732399 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732417 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732439 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv6zw\" (UniqueName: \"kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.732492 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.733540 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.736329 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.737197 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.739559 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.927927 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3576089d-8a48-4d20-9a43-acb95e60d037" path="/var/lib/kubelet/pods/3576089d-8a48-4d20-9a43-acb95e60d037/volumes" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.928952 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.929026 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:52 crc kubenswrapper[4867]: I1212 07:09:52.934209 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv6zw\" (UniqueName: \"kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw\") pod \"ceilometer-0\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " pod="openstack/ceilometer-0" Dec 12 07:09:53 crc kubenswrapper[4867]: I1212 07:09:53.111187 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:09:53 crc kubenswrapper[4867]: W1212 07:09:53.573503 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e785190_b6d8_494d_a2ad_d5ac5ca4e08b.slice/crio-d6057f692d50bb22029e58e37731107554ff76775aa6eddf265d34e5f50f0282 WatchSource:0}: Error finding container d6057f692d50bb22029e58e37731107554ff76775aa6eddf265d34e5f50f0282: Status 404 returned error can't find the container with id d6057f692d50bb22029e58e37731107554ff76775aa6eddf265d34e5f50f0282 Dec 12 07:09:53 crc kubenswrapper[4867]: I1212 07:09:53.577337 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:09:54 crc kubenswrapper[4867]: I1212 07:09:54.137559 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerStarted","Data":"d6057f692d50bb22029e58e37731107554ff76775aa6eddf265d34e5f50f0282"} Dec 12 07:09:55 crc kubenswrapper[4867]: I1212 07:09:55.147918 4867 generic.go:334] "Generic (PLEG): container finished" podID="714a3a25-e3f2-4847-9200-0c633bc0254f" containerID="2c02009ab1e1505ffea6c57acadcdd4d133ceb5623c0d8e451ec9680ea621aeb" exitCode=0 Dec 12 07:09:55 crc kubenswrapper[4867]: I1212 07:09:55.147999 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" event={"ID":"714a3a25-e3f2-4847-9200-0c633bc0254f","Type":"ContainerDied","Data":"2c02009ab1e1505ffea6c57acadcdd4d133ceb5623c0d8e451ec9680ea621aeb"} Dec 12 07:09:55 crc kubenswrapper[4867]: I1212 07:09:55.150118 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerStarted","Data":"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0"} Dec 12 07:09:55 crc kubenswrapper[4867]: I1212 07:09:55.150158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerStarted","Data":"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3"} Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.162263 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerStarted","Data":"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f"} Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.542454 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.713413 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data\") pod \"714a3a25-e3f2-4847-9200-0c633bc0254f\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.713503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6v7v\" (UniqueName: \"kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v\") pod \"714a3a25-e3f2-4847-9200-0c633bc0254f\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.713545 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts\") pod \"714a3a25-e3f2-4847-9200-0c633bc0254f\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.713684 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle\") pod \"714a3a25-e3f2-4847-9200-0c633bc0254f\" (UID: \"714a3a25-e3f2-4847-9200-0c633bc0254f\") " Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.720088 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v" (OuterVolumeSpecName: "kube-api-access-s6v7v") pod "714a3a25-e3f2-4847-9200-0c633bc0254f" (UID: "714a3a25-e3f2-4847-9200-0c633bc0254f"). InnerVolumeSpecName "kube-api-access-s6v7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.725795 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts" (OuterVolumeSpecName: "scripts") pod "714a3a25-e3f2-4847-9200-0c633bc0254f" (UID: "714a3a25-e3f2-4847-9200-0c633bc0254f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.740588 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "714a3a25-e3f2-4847-9200-0c633bc0254f" (UID: "714a3a25-e3f2-4847-9200-0c633bc0254f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.743446 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data" (OuterVolumeSpecName: "config-data") pod "714a3a25-e3f2-4847-9200-0c633bc0254f" (UID: "714a3a25-e3f2-4847-9200-0c633bc0254f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.827976 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.828033 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6v7v\" (UniqueName: \"kubernetes.io/projected/714a3a25-e3f2-4847-9200-0c633bc0254f-kube-api-access-s6v7v\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.828048 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:56 crc kubenswrapper[4867]: I1212 07:09:56.828059 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/714a3a25-e3f2-4847-9200-0c633bc0254f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.175152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" event={"ID":"714a3a25-e3f2-4847-9200-0c633bc0254f","Type":"ContainerDied","Data":"dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021"} Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.175452 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbae8110c7b702ef4b832770b54ff20248a456cfda6ca34dfc7b9d84817d8021" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.175208 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xkgvg" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.275182 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:09:57 crc kubenswrapper[4867]: E1212 07:09:57.275713 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714a3a25-e3f2-4847-9200-0c633bc0254f" containerName="nova-cell0-conductor-db-sync" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.275739 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="714a3a25-e3f2-4847-9200-0c633bc0254f" containerName="nova-cell0-conductor-db-sync" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.275997 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="714a3a25-e3f2-4847-9200-0c633bc0254f" containerName="nova-cell0-conductor-db-sync" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.276790 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.278788 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l679z" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.281252 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.310388 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.337972 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.338058 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.338103 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhtm7\" (UniqueName: \"kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.439668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhtm7\" (UniqueName: \"kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.439841 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.439879 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.445606 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.446298 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.457115 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhtm7\" (UniqueName: \"kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7\") pod \"nova-cell0-conductor-0\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:57 crc kubenswrapper[4867]: I1212 07:09:57.598452 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:58 crc kubenswrapper[4867]: I1212 07:09:58.057298 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:09:58 crc kubenswrapper[4867]: W1212 07:09:58.070868 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1867279_1908_42f2_aa75_b70707c44504.slice/crio-2fd04761dbcd97b8a31d03336970fba41ab75a88989af157e0de73767ca4276b WatchSource:0}: Error finding container 2fd04761dbcd97b8a31d03336970fba41ab75a88989af157e0de73767ca4276b: Status 404 returned error can't find the container with id 2fd04761dbcd97b8a31d03336970fba41ab75a88989af157e0de73767ca4276b Dec 12 07:09:58 crc kubenswrapper[4867]: I1212 07:09:58.241601 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerStarted","Data":"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e"} Dec 12 07:09:58 crc kubenswrapper[4867]: I1212 07:09:58.264830 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:09:58 crc kubenswrapper[4867]: I1212 07:09:58.273295 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f1867279-1908-42f2-aa75-b70707c44504","Type":"ContainerStarted","Data":"2fd04761dbcd97b8a31d03336970fba41ab75a88989af157e0de73767ca4276b"} Dec 12 07:09:58 crc kubenswrapper[4867]: I1212 07:09:58.303924 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.894139145 podStartE2EDuration="6.303898544s" podCreationTimestamp="2025-12-12 07:09:52 +0000 UTC" firstStartedPulling="2025-12-12 07:09:53.576250522 +0000 UTC m=+1281.147631791" lastFinishedPulling="2025-12-12 07:09:56.986009921 +0000 UTC m=+1284.557391190" observedRunningTime="2025-12-12 07:09:58.297062297 +0000 UTC m=+1285.868443586" watchObservedRunningTime="2025-12-12 07:09:58.303898544 +0000 UTC m=+1285.875279813" Dec 12 07:09:59 crc kubenswrapper[4867]: I1212 07:09:59.288262 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f1867279-1908-42f2-aa75-b70707c44504","Type":"ContainerStarted","Data":"19e718fb4f25322d032eec92be94441f173c338af8558ce522ff9f0484da2831"} Dec 12 07:09:59 crc kubenswrapper[4867]: I1212 07:09:59.288646 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 12 07:09:59 crc kubenswrapper[4867]: I1212 07:09:59.315246 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.315210968 podStartE2EDuration="2.315210968s" podCreationTimestamp="2025-12-12 07:09:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:09:59.305603121 +0000 UTC m=+1286.876984420" watchObservedRunningTime="2025-12-12 07:09:59.315210968 +0000 UTC m=+1286.886592237" Dec 12 07:10:00 crc kubenswrapper[4867]: I1212 07:10:00.919752 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:01 crc kubenswrapper[4867]: I1212 07:10:01.306167 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f1867279-1908-42f2-aa75-b70707c44504" containerName="nova-cell0-conductor-conductor" containerID="cri-o://19e718fb4f25322d032eec92be94441f173c338af8558ce522ff9f0484da2831" gracePeriod=30 Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.321912 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1867279-1908-42f2-aa75-b70707c44504" containerID="19e718fb4f25322d032eec92be94441f173c338af8558ce522ff9f0484da2831" exitCode=0 Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.321978 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f1867279-1908-42f2-aa75-b70707c44504","Type":"ContainerDied","Data":"19e718fb4f25322d032eec92be94441f173c338af8558ce522ff9f0484da2831"} Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.432035 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.460750 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data\") pod \"f1867279-1908-42f2-aa75-b70707c44504\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.461090 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhtm7\" (UniqueName: \"kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7\") pod \"f1867279-1908-42f2-aa75-b70707c44504\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.461339 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle\") pod \"f1867279-1908-42f2-aa75-b70707c44504\" (UID: \"f1867279-1908-42f2-aa75-b70707c44504\") " Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.478183 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7" (OuterVolumeSpecName: "kube-api-access-rhtm7") pod "f1867279-1908-42f2-aa75-b70707c44504" (UID: "f1867279-1908-42f2-aa75-b70707c44504"). InnerVolumeSpecName "kube-api-access-rhtm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.492249 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data" (OuterVolumeSpecName: "config-data") pod "f1867279-1908-42f2-aa75-b70707c44504" (UID: "f1867279-1908-42f2-aa75-b70707c44504"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.506730 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1867279-1908-42f2-aa75-b70707c44504" (UID: "f1867279-1908-42f2-aa75-b70707c44504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.563564 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.563595 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhtm7\" (UniqueName: \"kubernetes.io/projected/f1867279-1908-42f2-aa75-b70707c44504-kube-api-access-rhtm7\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.563608 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1867279-1908-42f2-aa75-b70707c44504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.993449 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.994027 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-central-agent" containerID="cri-o://d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" gracePeriod=30 Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.994125 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="proxy-httpd" containerID="cri-o://d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" gracePeriod=30 Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.994121 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="sg-core" containerID="cri-o://8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" gracePeriod=30 Dec 12 07:10:02 crc kubenswrapper[4867]: I1212 07:10:02.994125 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-notification-agent" containerID="cri-o://bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" gracePeriod=30 Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.335749 4867 generic.go:334] "Generic (PLEG): container finished" podID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerID="d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" exitCode=0 Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.335789 4867 generic.go:334] "Generic (PLEG): container finished" podID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerID="8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" exitCode=2 Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.335846 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerDied","Data":"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e"} Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.335879 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerDied","Data":"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f"} Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.338862 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f1867279-1908-42f2-aa75-b70707c44504","Type":"ContainerDied","Data":"2fd04761dbcd97b8a31d03336970fba41ab75a88989af157e0de73767ca4276b"} Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.338896 4867 scope.go:117] "RemoveContainer" containerID="19e718fb4f25322d032eec92be94441f173c338af8558ce522ff9f0484da2831" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.338950 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.378090 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.397340 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.405549 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:03 crc kubenswrapper[4867]: E1212 07:10:03.405943 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1867279-1908-42f2-aa75-b70707c44504" containerName="nova-cell0-conductor-conductor" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.405959 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1867279-1908-42f2-aa75-b70707c44504" containerName="nova-cell0-conductor-conductor" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.406161 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1867279-1908-42f2-aa75-b70707c44504" containerName="nova-cell0-conductor-conductor" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.406951 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.418511 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-l679z" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.419167 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.427800 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.478661 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.478759 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.478820 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9mn9\" (UniqueName: \"kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.580861 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.580973 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.581027 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9mn9\" (UniqueName: \"kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.588684 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.588696 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.599463 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9mn9\" (UniqueName: \"kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9\") pod \"nova-cell0-conductor-0\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.740650 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.783503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.783674 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.783725 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv6zw\" (UniqueName: \"kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.783804 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.784597 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.784843 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.784890 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.784909 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml\") pod \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\" (UID: \"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b\") " Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.784761 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.785579 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.785606 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.789133 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts" (OuterVolumeSpecName: "scripts") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.789301 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw" (OuterVolumeSpecName: "kube-api-access-qv6zw") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "kube-api-access-qv6zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.810836 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.857185 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.874179 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data" (OuterVolumeSpecName: "config-data") pod "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" (UID: "7e785190-b6d8-494d-a2ad-d5ac5ca4e08b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.877599 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.887120 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.887158 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.887171 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.887184 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:03 crc kubenswrapper[4867]: I1212 07:10:03.887194 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv6zw\" (UniqueName: \"kubernetes.io/projected/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b-kube-api-access-qv6zw\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.319801 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.353815 4867 generic.go:334] "Generic (PLEG): container finished" podID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerID="bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" exitCode=0 Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.354270 4867 generic.go:334] "Generic (PLEG): container finished" podID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerID="d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" exitCode=0 Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.354418 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.355302 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerDied","Data":"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0"} Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.355378 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerDied","Data":"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3"} Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.355396 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e785190-b6d8-494d-a2ad-d5ac5ca4e08b","Type":"ContainerDied","Data":"d6057f692d50bb22029e58e37731107554ff76775aa6eddf265d34e5f50f0282"} Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.355440 4867 scope.go:117] "RemoveContainer" containerID="d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.358414 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f","Type":"ContainerStarted","Data":"78efefccff02d40e18cef31a2da4391686812283e0342c0595524e040fbdeb99"} Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.458827 4867 scope.go:117] "RemoveContainer" containerID="8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.484330 4867 scope.go:117] "RemoveContainer" containerID="bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.485953 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.509309 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.516980 4867 scope.go:117] "RemoveContainer" containerID="d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.521353 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.521749 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-central-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.521765 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-central-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.521782 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="proxy-httpd" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.521788 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="proxy-httpd" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.521803 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="sg-core" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.521809 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="sg-core" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.521836 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-notification-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.521842 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-notification-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.522013 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="sg-core" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.522028 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="proxy-httpd" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.522047 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-central-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.522057 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" containerName="ceilometer-notification-agent" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.523603 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.527111 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.527317 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.530716 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545464 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545509 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545528 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545606 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545628 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545654 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.545704 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkd6d\" (UniqueName: \"kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.557280 4867 scope.go:117] "RemoveContainer" containerID="d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.557907 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e\": container with ID starting with d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e not found: ID does not exist" containerID="d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.557949 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e"} err="failed to get container status \"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e\": rpc error: code = NotFound desc = could not find container \"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e\": container with ID starting with d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.557976 4867 scope.go:117] "RemoveContainer" containerID="8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.558637 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f\": container with ID starting with 8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f not found: ID does not exist" containerID="8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.558667 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f"} err="failed to get container status \"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f\": rpc error: code = NotFound desc = could not find container \"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f\": container with ID starting with 8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.558686 4867 scope.go:117] "RemoveContainer" containerID="bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.559050 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0\": container with ID starting with bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0 not found: ID does not exist" containerID="bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559074 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0"} err="failed to get container status \"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0\": rpc error: code = NotFound desc = could not find container \"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0\": container with ID starting with bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0 not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559090 4867 scope.go:117] "RemoveContainer" containerID="d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" Dec 12 07:10:04 crc kubenswrapper[4867]: E1212 07:10:04.559412 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3\": container with ID starting with d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3 not found: ID does not exist" containerID="d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559438 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3"} err="failed to get container status \"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3\": rpc error: code = NotFound desc = could not find container \"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3\": container with ID starting with d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3 not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559457 4867 scope.go:117] "RemoveContainer" containerID="d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559812 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e"} err="failed to get container status \"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e\": rpc error: code = NotFound desc = could not find container \"d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e\": container with ID starting with d700301ad1190d3ad5b63a7394650102abafbcba354e9fb1a33c35fdf72ff92e not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.559836 4867 scope.go:117] "RemoveContainer" containerID="8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.560084 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f"} err="failed to get container status \"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f\": rpc error: code = NotFound desc = could not find container \"8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f\": container with ID starting with 8c08863c95d0cb46190ab233ba5cefebdcc5339254e7afbcac1163351d2a3b6f not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.560107 4867 scope.go:117] "RemoveContainer" containerID="bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.560387 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0"} err="failed to get container status \"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0\": rpc error: code = NotFound desc = could not find container \"bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0\": container with ID starting with bcecdc8358a841524cf4aa257575c63b0bf1b6879652c8d817c7f07d6494c5d0 not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.560418 4867 scope.go:117] "RemoveContainer" containerID="d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.560611 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3"} err="failed to get container status \"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3\": rpc error: code = NotFound desc = could not find container \"d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3\": container with ID starting with d2d3db9a857371f66faec059eea32d1fd8659ff2f642a47df16381011be08dc3 not found: ID does not exist" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.647567 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkd6d\" (UniqueName: \"kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.647705 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.649455 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.649528 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.649716 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.649769 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.649837 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.650470 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.650518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.657516 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.662012 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.662183 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.662650 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.681123 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkd6d\" (UniqueName: \"kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d\") pod \"ceilometer-0\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.850466 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.864290 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e785190-b6d8-494d-a2ad-d5ac5ca4e08b" path="/var/lib/kubelet/pods/7e785190-b6d8-494d-a2ad-d5ac5ca4e08b/volumes" Dec 12 07:10:04 crc kubenswrapper[4867]: I1212 07:10:04.866452 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1867279-1908-42f2-aa75-b70707c44504" path="/var/lib/kubelet/pods/f1867279-1908-42f2-aa75-b70707c44504/volumes" Dec 12 07:10:05 crc kubenswrapper[4867]: I1212 07:10:05.317157 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:05 crc kubenswrapper[4867]: I1212 07:10:05.369858 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f","Type":"ContainerStarted","Data":"848266a1cfe8cee63f4a2009ad41e8a49241ba1784825fa11ad76c5f036dc5c0"} Dec 12 07:10:05 crc kubenswrapper[4867]: I1212 07:10:05.371121 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:05 crc kubenswrapper[4867]: I1212 07:10:05.373996 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerStarted","Data":"f98297e7f805ae8effb2c1dab9f7b5595a042caab60f480043966d9642293c30"} Dec 12 07:10:05 crc kubenswrapper[4867]: I1212 07:10:05.397611 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.397590459 podStartE2EDuration="2.397590459s" podCreationTimestamp="2025-12-12 07:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:05.386201669 +0000 UTC m=+1292.957582948" watchObservedRunningTime="2025-12-12 07:10:05.397590459 +0000 UTC m=+1292.968971728" Dec 12 07:10:06 crc kubenswrapper[4867]: I1212 07:10:06.387849 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerStarted","Data":"0e9ae4f70c87c4971daedb51b9b5e5e54d2785732a42907c01159ee86492c662"} Dec 12 07:10:07 crc kubenswrapper[4867]: I1212 07:10:07.397680 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerStarted","Data":"9362ae3393372fcca9fb0c7bb91660a1a6e4882a5e215ffb2b52e9f6c1bff4cb"} Dec 12 07:10:08 crc kubenswrapper[4867]: I1212 07:10:08.410561 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerStarted","Data":"6328805576d443af9afecb5b9296d72aa88d48e9f0c4504a04073e33616ff1fa"} Dec 12 07:10:10 crc kubenswrapper[4867]: I1212 07:10:10.428447 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerStarted","Data":"edac33da98bf50816eab183b68a6de9858d9eaa542a52c8b09931db9f11a16d1"} Dec 12 07:10:10 crc kubenswrapper[4867]: I1212 07:10:10.429004 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:10:10 crc kubenswrapper[4867]: I1212 07:10:10.453668 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.519781661 podStartE2EDuration="6.453650203s" podCreationTimestamp="2025-12-12 07:10:04 +0000 UTC" firstStartedPulling="2025-12-12 07:10:05.32516858 +0000 UTC m=+1292.896549849" lastFinishedPulling="2025-12-12 07:10:09.259037122 +0000 UTC m=+1296.830418391" observedRunningTime="2025-12-12 07:10:10.446370792 +0000 UTC m=+1298.017752071" watchObservedRunningTime="2025-12-12 07:10:10.453650203 +0000 UTC m=+1298.025031472" Dec 12 07:10:13 crc kubenswrapper[4867]: I1212 07:10:13.909912 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.471198 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zjmsk"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.474929 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.477368 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.478646 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.487978 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zjmsk"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.643875 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.644336 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.644490 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cws2s\" (UniqueName: \"kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.644527 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.672110 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.673704 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.680927 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.684957 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.746602 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cws2s\" (UniqueName: \"kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.746668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.746720 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.746811 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.756430 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.758253 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.768032 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.778970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cws2s\" (UniqueName: \"kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s\") pod \"nova-cell0-cell-mapping-zjmsk\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.798411 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.802887 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.804754 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.810402 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.827023 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.848341 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.848394 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.848414 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.848481 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5948\" (UniqueName: \"kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951256 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951371 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951396 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grdmg\" (UniqueName: \"kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951420 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951436 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951484 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.951519 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5948\" (UniqueName: \"kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.953217 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:14 crc kubenswrapper[4867]: I1212 07:10:14.959887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.013323 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5948\" (UniqueName: \"kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.030587 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " pod="openstack/nova-api-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.053885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grdmg\" (UniqueName: \"kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.053981 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.054056 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.056993 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.076912 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.091484 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.094843 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.095833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grdmg\" (UniqueName: \"kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg\") pod \"nova-scheduler-0\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.104100 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.144753 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.203478 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.205038 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.211907 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.218939 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.244843 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.246080 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.247718 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.258557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.258676 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.258717 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.258743 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbdgw\" (UniqueName: \"kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.258841 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.293320 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360709 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360793 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360871 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360942 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360965 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.360989 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j9g2\" (UniqueName: \"kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361014 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbdgw\" (UniqueName: \"kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361118 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361149 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361179 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361215 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptlg8\" (UniqueName: \"kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.361217 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.365203 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.365802 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.382997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbdgw\" (UniqueName: \"kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw\") pod \"nova-metadata-0\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463120 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463499 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptlg8\" (UniqueName: \"kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463549 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463791 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463835 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463863 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j9g2\" (UniqueName: \"kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463954 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.463975 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.466194 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.467742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.468607 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.468631 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.469191 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.471580 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.478822 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.496025 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.503820 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptlg8\" (UniqueName: \"kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8\") pod \"dnsmasq-dns-647df7b8c5-47dxv\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.506166 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j9g2\" (UniqueName: \"kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2\") pod \"nova-cell1-novncproxy-0\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.545728 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.582265 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.583570 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zjmsk"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.773562 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dlfmg"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.775298 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.780784 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.780900 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.797465 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dlfmg"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.873549 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.873625 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.873655 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqjm4\" (UniqueName: \"kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.873847 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.873977 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.922841 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.978540 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.978990 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.979021 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqjm4\" (UniqueName: \"kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.979174 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.984422 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:15 crc kubenswrapper[4867]: I1212 07:10:15.988998 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.017997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.020061 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.028867 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqjm4\" (UniqueName: \"kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4\") pod \"nova-cell1-conductor-db-sync-dlfmg\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.126029 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.152990 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:16 crc kubenswrapper[4867]: W1212 07:10:16.162597 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a7cf5a6_b0d7_404c_afe2_c4717fb0eeb6.slice/crio-a41ca706c58ff4d8914beef84dda3cdf3ff8dc8e277d28916c54517fbfb216e3 WatchSource:0}: Error finding container a41ca706c58ff4d8914beef84dda3cdf3ff8dc8e277d28916c54517fbfb216e3: Status 404 returned error can't find the container with id a41ca706c58ff4d8914beef84dda3cdf3ff8dc8e277d28916c54517fbfb216e3 Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.417373 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:10:16 crc kubenswrapper[4867]: W1212 07:10:16.432635 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2dfa393_ab5a_4aad_b7cc_cdc49d3d1da9.slice/crio-757a4b1d9238758b4288cf09717a12ffc0850f78a43bda807275549fe26be2e7 WatchSource:0}: Error finding container 757a4b1d9238758b4288cf09717a12ffc0850f78a43bda807275549fe26be2e7: Status 404 returned error can't find the container with id 757a4b1d9238758b4288cf09717a12ffc0850f78a43bda807275549fe26be2e7 Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.444188 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.492760 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" event={"ID":"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9","Type":"ContainerStarted","Data":"757a4b1d9238758b4288cf09717a12ffc0850f78a43bda807275549fe26be2e7"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.494780 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd0ec4ba-7992-4650-a749-5fd94190938d","Type":"ContainerStarted","Data":"aa590cc23fc2b3a641f03ed11798ec7488deb7992789614faffd7122cbf49b1c"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.497499 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e0a447bc-40d5-401c-80c8-14d5d5385805","Type":"ContainerStarted","Data":"0ca5b74d791cb2896066fe7d370394b5a862d406544245447f4dcfc041d7605d"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.500664 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerStarted","Data":"a41ca706c58ff4d8914beef84dda3cdf3ff8dc8e277d28916c54517fbfb216e3"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.502511 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerStarted","Data":"46a3ea400c92dfe1e5ea9101fe63b45fd9f0a1ea2b3fef8c230395f3f20eba4b"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.508647 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zjmsk" event={"ID":"469654c1-0303-4127-a61c-411d65a09cd2","Type":"ContainerStarted","Data":"dc7c6abeff19f2ed7d1082efbb635045ee7b036f7c1606323782742b6b4993b6"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.508706 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zjmsk" event={"ID":"469654c1-0303-4127-a61c-411d65a09cd2","Type":"ContainerStarted","Data":"8e6c96ffa4cd97eccd0eef860e01c4aafb39e398a646f56b18fb3c80410c428f"} Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.571161 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zjmsk" podStartSLOduration=2.571140045 podStartE2EDuration="2.571140045s" podCreationTimestamp="2025-12-12 07:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:16.565816743 +0000 UTC m=+1304.137198012" watchObservedRunningTime="2025-12-12 07:10:16.571140045 +0000 UTC m=+1304.142521344" Dec 12 07:10:16 crc kubenswrapper[4867]: I1212 07:10:16.695170 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dlfmg"] Dec 12 07:10:17 crc kubenswrapper[4867]: I1212 07:10:17.519120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" event={"ID":"44b73a04-c864-44ed-b88d-9b6a5073861a","Type":"ContainerStarted","Data":"1eb349ad10dac75fbf2979864d21a0ab5fd4b851e81f9bfe08f8809cf9cf72f8"} Dec 12 07:10:17 crc kubenswrapper[4867]: I1212 07:10:17.519781 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" event={"ID":"44b73a04-c864-44ed-b88d-9b6a5073861a","Type":"ContainerStarted","Data":"a7a5b947e5e8bdac366d6c7683882c037750719bd78663be978163aae7f9e718"} Dec 12 07:10:17 crc kubenswrapper[4867]: I1212 07:10:17.523664 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerID="09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58" exitCode=0 Dec 12 07:10:17 crc kubenswrapper[4867]: I1212 07:10:17.523760 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" event={"ID":"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9","Type":"ContainerDied","Data":"09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58"} Dec 12 07:10:17 crc kubenswrapper[4867]: I1212 07:10:17.568148 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" podStartSLOduration=2.5681258959999997 podStartE2EDuration="2.568125896s" podCreationTimestamp="2025-12-12 07:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:17.544671934 +0000 UTC m=+1305.116053203" watchObservedRunningTime="2025-12-12 07:10:17.568125896 +0000 UTC m=+1305.139507165" Dec 12 07:10:18 crc kubenswrapper[4867]: I1212 07:10:18.537015 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" event={"ID":"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9","Type":"ContainerStarted","Data":"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c"} Dec 12 07:10:18 crc kubenswrapper[4867]: I1212 07:10:18.537416 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:18 crc kubenswrapper[4867]: I1212 07:10:18.560984 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" podStartSLOduration=3.560958603 podStartE2EDuration="3.560958603s" podCreationTimestamp="2025-12-12 07:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:18.558319938 +0000 UTC m=+1306.129701197" watchObservedRunningTime="2025-12-12 07:10:18.560958603 +0000 UTC m=+1306.132339872" Dec 12 07:10:18 crc kubenswrapper[4867]: I1212 07:10:18.621013 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:18 crc kubenswrapper[4867]: I1212 07:10:18.639896 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.588968 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd0ec4ba-7992-4650-a749-5fd94190938d","Type":"ContainerStarted","Data":"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.589091 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fd0ec4ba-7992-4650-a749-5fd94190938d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8" gracePeriod=30 Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.593301 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e0a447bc-40d5-401c-80c8-14d5d5385805","Type":"ContainerStarted","Data":"6b73343f07d8e0ffe2561710edb199fe0ff3942f477bcbaa5d42a9ebf85bc561"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.595555 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerStarted","Data":"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.595592 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerStarted","Data":"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.595650 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-log" containerID="cri-o://7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" gracePeriod=30 Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.595712 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-metadata" containerID="cri-o://acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" gracePeriod=30 Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.598023 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerStarted","Data":"f2d06888e1824b50ccb88e51caad856d195a7d84b98146b6a12ccf98f22cdf19"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.598057 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerStarted","Data":"d7fd8d67fa6de3d31075f82dfc86b3e76c705c975db03a4058b3357edc6f3b2a"} Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.613815 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.5099908319999997 podStartE2EDuration="8.613790839s" podCreationTimestamp="2025-12-12 07:10:14 +0000 UTC" firstStartedPulling="2025-12-12 07:10:16.452747324 +0000 UTC m=+1304.024128593" lastFinishedPulling="2025-12-12 07:10:21.556547331 +0000 UTC m=+1309.127928600" observedRunningTime="2025-12-12 07:10:22.605245467 +0000 UTC m=+1310.176626756" watchObservedRunningTime="2025-12-12 07:10:22.613790839 +0000 UTC m=+1310.185172108" Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.627864 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.278530541 podStartE2EDuration="8.627841558s" podCreationTimestamp="2025-12-12 07:10:14 +0000 UTC" firstStartedPulling="2025-12-12 07:10:16.165466036 +0000 UTC m=+1303.736847305" lastFinishedPulling="2025-12-12 07:10:21.514777053 +0000 UTC m=+1309.086158322" observedRunningTime="2025-12-12 07:10:22.624960566 +0000 UTC m=+1310.196341845" watchObservedRunningTime="2025-12-12 07:10:22.627841558 +0000 UTC m=+1310.199222827" Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.642814 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.050001504 podStartE2EDuration="8.64279065s" podCreationTimestamp="2025-12-12 07:10:14 +0000 UTC" firstStartedPulling="2025-12-12 07:10:15.922596802 +0000 UTC m=+1303.493978071" lastFinishedPulling="2025-12-12 07:10:21.515385958 +0000 UTC m=+1309.086767217" observedRunningTime="2025-12-12 07:10:22.639463906 +0000 UTC m=+1310.210845175" watchObservedRunningTime="2025-12-12 07:10:22.64279065 +0000 UTC m=+1310.214171919" Dec 12 07:10:22 crc kubenswrapper[4867]: I1212 07:10:22.659184 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.174873445 podStartE2EDuration="8.659161656s" podCreationTimestamp="2025-12-12 07:10:14 +0000 UTC" firstStartedPulling="2025-12-12 07:10:16.033549568 +0000 UTC m=+1303.604930837" lastFinishedPulling="2025-12-12 07:10:21.517837769 +0000 UTC m=+1309.089219048" observedRunningTime="2025-12-12 07:10:22.656697385 +0000 UTC m=+1310.228078654" watchObservedRunningTime="2025-12-12 07:10:22.659161656 +0000 UTC m=+1310.230542935" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.271476 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.360831 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbdgw\" (UniqueName: \"kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw\") pod \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.361028 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data\") pod \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.361072 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs\") pod \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.361138 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle\") pod \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\" (UID: \"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6\") " Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.364512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs" (OuterVolumeSpecName: "logs") pod "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" (UID: "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.376792 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw" (OuterVolumeSpecName: "kube-api-access-bbdgw") pod "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" (UID: "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6"). InnerVolumeSpecName "kube-api-access-bbdgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.406418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" (UID: "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.417776 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data" (OuterVolumeSpecName: "config-data") pod "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" (UID: "8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.463136 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.463163 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.463171 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.463182 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbdgw\" (UniqueName: \"kubernetes.io/projected/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6-kube-api-access-bbdgw\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609081 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerID="acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" exitCode=0 Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609125 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerID="7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" exitCode=143 Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609554 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609548 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerDied","Data":"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81"} Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609699 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerDied","Data":"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439"} Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6","Type":"ContainerDied","Data":"a41ca706c58ff4d8914beef84dda3cdf3ff8dc8e277d28916c54517fbfb216e3"} Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.609772 4867 scope.go:117] "RemoveContainer" containerID="acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.636692 4867 scope.go:117] "RemoveContainer" containerID="7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.646727 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.670936 4867 scope.go:117] "RemoveContainer" containerID="acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" Dec 12 07:10:23 crc kubenswrapper[4867]: E1212 07:10:23.676614 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81\": container with ID starting with acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81 not found: ID does not exist" containerID="acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.676673 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81"} err="failed to get container status \"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81\": rpc error: code = NotFound desc = could not find container \"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81\": container with ID starting with acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81 not found: ID does not exist" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.676765 4867 scope.go:117] "RemoveContainer" containerID="7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" Dec 12 07:10:23 crc kubenswrapper[4867]: E1212 07:10:23.678406 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439\": container with ID starting with 7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439 not found: ID does not exist" containerID="7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.678446 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439"} err="failed to get container status \"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439\": rpc error: code = NotFound desc = could not find container \"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439\": container with ID starting with 7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439 not found: ID does not exist" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.678469 4867 scope.go:117] "RemoveContainer" containerID="acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.678530 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.678809 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81"} err="failed to get container status \"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81\": rpc error: code = NotFound desc = could not find container \"acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81\": container with ID starting with acee330f872fcb956c755d99c0a019ad14471213a698d426e64d1b99b423ee81 not found: ID does not exist" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.678854 4867 scope.go:117] "RemoveContainer" containerID="7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.682382 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439"} err="failed to get container status \"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439\": rpc error: code = NotFound desc = could not find container \"7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439\": container with ID starting with 7bfdcdf00bd9d44420ce6b6f5a860dc62e231fb96bdab4ea5890ad62b679b439 not found: ID does not exist" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.689291 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:23 crc kubenswrapper[4867]: E1212 07:10:23.690044 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-log" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.690160 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-log" Dec 12 07:10:23 crc kubenswrapper[4867]: E1212 07:10:23.690324 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-metadata" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.690406 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-metadata" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.690721 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-log" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.690822 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" containerName="nova-metadata-metadata" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.692198 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.697099 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.697523 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.697628 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.779632 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h92cw\" (UniqueName: \"kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.779701 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.779875 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.779897 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.779937 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882007 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882059 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882094 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882143 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h92cw\" (UniqueName: \"kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882206 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.882717 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.888502 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.889002 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.891922 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:23 crc kubenswrapper[4867]: I1212 07:10:23.918930 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h92cw\" (UniqueName: \"kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw\") pod \"nova-metadata-0\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " pod="openstack/nova-metadata-0" Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.013078 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.437280 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.624555 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerStarted","Data":"455c4652f6ec533eb04e74d0e91a497137c298fa66e3a5cd1e1a65bd485a7507"} Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.626134 4867 generic.go:334] "Generic (PLEG): container finished" podID="469654c1-0303-4127-a61c-411d65a09cd2" containerID="dc7c6abeff19f2ed7d1082efbb635045ee7b036f7c1606323782742b6b4993b6" exitCode=0 Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.626188 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zjmsk" event={"ID":"469654c1-0303-4127-a61c-411d65a09cd2","Type":"ContainerDied","Data":"dc7c6abeff19f2ed7d1082efbb635045ee7b036f7c1606323782742b6b4993b6"} Dec 12 07:10:24 crc kubenswrapper[4867]: I1212 07:10:24.855273 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6" path="/var/lib/kubelet/pods/8a7cf5a6-b0d7-404c-afe2-c4717fb0eeb6/volumes" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.245195 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.245550 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.272051 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.293677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.293737 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.548076 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.583457 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.645343 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerStarted","Data":"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596"} Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.645409 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerStarted","Data":"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853"} Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.666644 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.666892 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="dnsmasq-dns" containerID="cri-o://ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc" gracePeriod=10 Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.678695 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.678675018 podStartE2EDuration="2.678675018s" podCreationTimestamp="2025-12-12 07:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:25.670251889 +0000 UTC m=+1313.241633158" watchObservedRunningTime="2025-12-12 07:10:25.678675018 +0000 UTC m=+1313.250056287" Dec 12 07:10:25 crc kubenswrapper[4867]: I1212 07:10:25.744931 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.141436 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.227970 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cws2s\" (UniqueName: \"kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s\") pod \"469654c1-0303-4127-a61c-411d65a09cd2\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.228047 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts\") pod \"469654c1-0303-4127-a61c-411d65a09cd2\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.228077 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle\") pod \"469654c1-0303-4127-a61c-411d65a09cd2\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.228111 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data\") pod \"469654c1-0303-4127-a61c-411d65a09cd2\" (UID: \"469654c1-0303-4127-a61c-411d65a09cd2\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.247886 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s" (OuterVolumeSpecName: "kube-api-access-cws2s") pod "469654c1-0303-4127-a61c-411d65a09cd2" (UID: "469654c1-0303-4127-a61c-411d65a09cd2"). InnerVolumeSpecName "kube-api-access-cws2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.250665 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts" (OuterVolumeSpecName: "scripts") pod "469654c1-0303-4127-a61c-411d65a09cd2" (UID: "469654c1-0303-4127-a61c-411d65a09cd2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.268795 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "469654c1-0303-4127-a61c-411d65a09cd2" (UID: "469654c1-0303-4127-a61c-411d65a09cd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.291766 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data" (OuterVolumeSpecName: "config-data") pod "469654c1-0303-4127-a61c-411d65a09cd2" (UID: "469654c1-0303-4127-a61c-411d65a09cd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.330721 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cws2s\" (UniqueName: \"kubernetes.io/projected/469654c1-0303-4127-a61c-411d65a09cd2-kube-api-access-cws2s\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.330754 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.330768 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.330781 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/469654c1-0303-4127-a61c-411d65a09cd2-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.359999 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.381339 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.381629 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.182:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.434301 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lskb\" (UniqueName: \"kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.435080 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.435880 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.436073 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.436889 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.437391 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config\") pod \"6e6579f1-6cee-40dc-b1b1-fa309701574d\" (UID: \"6e6579f1-6cee-40dc-b1b1-fa309701574d\") " Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.443837 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb" (OuterVolumeSpecName: "kube-api-access-9lskb") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "kube-api-access-9lskb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.496596 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.501175 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.504938 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config" (OuterVolumeSpecName: "config") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.507990 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.513804 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e6579f1-6cee-40dc-b1b1-fa309701574d" (UID: "6e6579f1-6cee-40dc-b1b1-fa309701574d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541045 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541082 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541093 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541106 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lskb\" (UniqueName: \"kubernetes.io/projected/6e6579f1-6cee-40dc-b1b1-fa309701574d-kube-api-access-9lskb\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541116 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.541125 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6579f1-6cee-40dc-b1b1-fa309701574d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.654188 4867 generic.go:334] "Generic (PLEG): container finished" podID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerID="ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc" exitCode=0 Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.654260 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" event={"ID":"6e6579f1-6cee-40dc-b1b1-fa309701574d","Type":"ContainerDied","Data":"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc"} Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.654298 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" event={"ID":"6e6579f1-6cee-40dc-b1b1-fa309701574d","Type":"ContainerDied","Data":"10e261a2f434f43d7b0e451c7b4e80c28dea9a8eb57c9abd66a691ac2b54efb9"} Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.654320 4867 scope.go:117] "RemoveContainer" containerID="ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.654331 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dbb546bf-dqprf" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.656203 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zjmsk" event={"ID":"469654c1-0303-4127-a61c-411d65a09cd2","Type":"ContainerDied","Data":"8e6c96ffa4cd97eccd0eef860e01c4aafb39e398a646f56b18fb3c80410c428f"} Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.656254 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e6c96ffa4cd97eccd0eef860e01c4aafb39e398a646f56b18fb3c80410c428f" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.656330 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zjmsk" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.684322 4867 scope.go:117] "RemoveContainer" containerID="1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.712206 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.712305 4867 scope.go:117] "RemoveContainer" containerID="ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc" Dec 12 07:10:26 crc kubenswrapper[4867]: E1212 07:10:26.712631 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc\": container with ID starting with ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc not found: ID does not exist" containerID="ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.712665 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc"} err="failed to get container status \"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc\": rpc error: code = NotFound desc = could not find container \"ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc\": container with ID starting with ce6e30fdcc7ecde9b7952f8fe8b7dab5e1e0ebf57f8457a48b11aabcb24c4ecc not found: ID does not exist" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.712689 4867 scope.go:117] "RemoveContainer" containerID="1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6" Dec 12 07:10:26 crc kubenswrapper[4867]: E1212 07:10:26.712910 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6\": container with ID starting with 1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6 not found: ID does not exist" containerID="1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.712934 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6"} err="failed to get container status \"1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6\": rpc error: code = NotFound desc = could not find container \"1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6\": container with ID starting with 1f444e16feaab2a523292ab909449568ca1dcf25617a054e5ef82b45532365f6 not found: ID does not exist" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.718447 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75dbb546bf-dqprf"] Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.780160 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.780398 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-log" containerID="cri-o://d7fd8d67fa6de3d31075f82dfc86b3e76c705c975db03a4058b3357edc6f3b2a" gracePeriod=30 Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.780459 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-api" containerID="cri-o://f2d06888e1824b50ccb88e51caad856d195a7d84b98146b6a12ccf98f22cdf19" gracePeriod=30 Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.836284 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.862730 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" path="/var/lib/kubelet/pods/6e6579f1-6cee-40dc-b1b1-fa309701574d/volumes" Dec 12 07:10:26 crc kubenswrapper[4867]: I1212 07:10:26.863343 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:27 crc kubenswrapper[4867]: I1212 07:10:27.669367 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerID="d7fd8d67fa6de3d31075f82dfc86b3e76c705c975db03a4058b3357edc6f3b2a" exitCode=143 Dec 12 07:10:27 crc kubenswrapper[4867]: I1212 07:10:27.669447 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerDied","Data":"d7fd8d67fa6de3d31075f82dfc86b3e76c705c975db03a4058b3357edc6f3b2a"} Dec 12 07:10:27 crc kubenswrapper[4867]: I1212 07:10:27.670174 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e0a447bc-40d5-401c-80c8-14d5d5385805" containerName="nova-scheduler-scheduler" containerID="cri-o://6b73343f07d8e0ffe2561710edb199fe0ff3942f477bcbaa5d42a9ebf85bc561" gracePeriod=30 Dec 12 07:10:28 crc kubenswrapper[4867]: I1212 07:10:28.680305 4867 generic.go:334] "Generic (PLEG): container finished" podID="44b73a04-c864-44ed-b88d-9b6a5073861a" containerID="1eb349ad10dac75fbf2979864d21a0ab5fd4b851e81f9bfe08f8809cf9cf72f8" exitCode=0 Dec 12 07:10:28 crc kubenswrapper[4867]: I1212 07:10:28.680341 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" event={"ID":"44b73a04-c864-44ed-b88d-9b6a5073861a","Type":"ContainerDied","Data":"1eb349ad10dac75fbf2979864d21a0ab5fd4b851e81f9bfe08f8809cf9cf72f8"} Dec 12 07:10:28 crc kubenswrapper[4867]: I1212 07:10:28.680514 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-log" containerID="cri-o://4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" gracePeriod=30 Dec 12 07:10:28 crc kubenswrapper[4867]: I1212 07:10:28.680544 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-metadata" containerID="cri-o://e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" gracePeriod=30 Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.014202 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.014564 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.273370 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.399575 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h92cw\" (UniqueName: \"kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw\") pod \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.399624 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs\") pod \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.399828 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data\") pod \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.399922 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle\") pod \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.399955 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs\") pod \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\" (UID: \"e1d334ec-a346-4490-80b3-9a88bc2b91e7\") " Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.400699 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs" (OuterVolumeSpecName: "logs") pod "e1d334ec-a346-4490-80b3-9a88bc2b91e7" (UID: "e1d334ec-a346-4490-80b3-9a88bc2b91e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.405293 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw" (OuterVolumeSpecName: "kube-api-access-h92cw") pod "e1d334ec-a346-4490-80b3-9a88bc2b91e7" (UID: "e1d334ec-a346-4490-80b3-9a88bc2b91e7"). InnerVolumeSpecName "kube-api-access-h92cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.426480 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data" (OuterVolumeSpecName: "config-data") pod "e1d334ec-a346-4490-80b3-9a88bc2b91e7" (UID: "e1d334ec-a346-4490-80b3-9a88bc2b91e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.429881 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1d334ec-a346-4490-80b3-9a88bc2b91e7" (UID: "e1d334ec-a346-4490-80b3-9a88bc2b91e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.454069 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e1d334ec-a346-4490-80b3-9a88bc2b91e7" (UID: "e1d334ec-a346-4490-80b3-9a88bc2b91e7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.502045 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.502416 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.502505 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1d334ec-a346-4490-80b3-9a88bc2b91e7-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.502577 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h92cw\" (UniqueName: \"kubernetes.io/projected/e1d334ec-a346-4490-80b3-9a88bc2b91e7-kube-api-access-h92cw\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.502713 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1d334ec-a346-4490-80b3-9a88bc2b91e7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.692214 4867 generic.go:334] "Generic (PLEG): container finished" podID="e0a447bc-40d5-401c-80c8-14d5d5385805" containerID="6b73343f07d8e0ffe2561710edb199fe0ff3942f477bcbaa5d42a9ebf85bc561" exitCode=0 Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.692301 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e0a447bc-40d5-401c-80c8-14d5d5385805","Type":"ContainerDied","Data":"6b73343f07d8e0ffe2561710edb199fe0ff3942f477bcbaa5d42a9ebf85bc561"} Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.696707 4867 generic.go:334] "Generic (PLEG): container finished" podID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerID="e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" exitCode=0 Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.696827 4867 generic.go:334] "Generic (PLEG): container finished" podID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerID="4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" exitCode=143 Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.696758 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerDied","Data":"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596"} Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.697013 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerDied","Data":"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853"} Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.696761 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.697044 4867 scope.go:117] "RemoveContainer" containerID="e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.697034 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1d334ec-a346-4490-80b3-9a88bc2b91e7","Type":"ContainerDied","Data":"455c4652f6ec533eb04e74d0e91a497137c298fa66e3a5cd1e1a65bd485a7507"} Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.766568 4867 scope.go:117] "RemoveContainer" containerID="4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.773603 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.789520 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.816795 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.817522 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-log" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817547 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-log" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.817566 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-metadata" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817576 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-metadata" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.817605 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469654c1-0303-4127-a61c-411d65a09cd2" containerName="nova-manage" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817613 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="469654c1-0303-4127-a61c-411d65a09cd2" containerName="nova-manage" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.817626 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="dnsmasq-dns" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817633 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="dnsmasq-dns" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.817662 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="init" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817670 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="init" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817902 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-metadata" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817920 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" containerName="nova-metadata-log" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817934 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="469654c1-0303-4127-a61c-411d65a09cd2" containerName="nova-manage" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.817957 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6579f1-6cee-40dc-b1b1-fa309701574d" containerName="dnsmasq-dns" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.819365 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.829944 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.836779 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.837010 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.861631 4867 scope.go:117] "RemoveContainer" containerID="e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.862047 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596\": container with ID starting with e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596 not found: ID does not exist" containerID="e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862080 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596"} err="failed to get container status \"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596\": rpc error: code = NotFound desc = could not find container \"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596\": container with ID starting with e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596 not found: ID does not exist" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862105 4867 scope.go:117] "RemoveContainer" containerID="4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" Dec 12 07:10:29 crc kubenswrapper[4867]: E1212 07:10:29.862478 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853\": container with ID starting with 4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853 not found: ID does not exist" containerID="4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862576 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853"} err="failed to get container status \"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853\": rpc error: code = NotFound desc = could not find container \"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853\": container with ID starting with 4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853 not found: ID does not exist" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862634 4867 scope.go:117] "RemoveContainer" containerID="e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862916 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596"} err="failed to get container status \"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596\": rpc error: code = NotFound desc = could not find container \"e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596\": container with ID starting with e8f2485878a8115cf6c946d35d6b04610433acb61236284d5a29bad3527fe596 not found: ID does not exist" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.862943 4867 scope.go:117] "RemoveContainer" containerID="4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853" Dec 12 07:10:29 crc kubenswrapper[4867]: I1212 07:10:29.863322 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853"} err="failed to get container status \"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853\": rpc error: code = NotFound desc = could not find container \"4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853\": container with ID starting with 4fc8a454f001df15f6899ffb339a4dd29b8e25ebdcdba0d7cf4213e941c6e853 not found: ID does not exist" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.012348 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbpgx\" (UniqueName: \"kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.012409 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.012496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.012795 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.012834 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114350 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114468 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114514 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbpgx\" (UniqueName: \"kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114534 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114580 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.114810 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.119393 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.119419 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.129993 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.133219 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbpgx\" (UniqueName: \"kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx\") pod \"nova-metadata-0\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.165452 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.189581 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.196840 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.319867 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data\") pod \"44b73a04-c864-44ed-b88d-9b6a5073861a\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320391 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grdmg\" (UniqueName: \"kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg\") pod \"e0a447bc-40d5-401c-80c8-14d5d5385805\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320501 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqjm4\" (UniqueName: \"kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4\") pod \"44b73a04-c864-44ed-b88d-9b6a5073861a\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320538 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle\") pod \"e0a447bc-40d5-401c-80c8-14d5d5385805\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320567 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data\") pod \"e0a447bc-40d5-401c-80c8-14d5d5385805\" (UID: \"e0a447bc-40d5-401c-80c8-14d5d5385805\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320648 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts\") pod \"44b73a04-c864-44ed-b88d-9b6a5073861a\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.320689 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle\") pod \"44b73a04-c864-44ed-b88d-9b6a5073861a\" (UID: \"44b73a04-c864-44ed-b88d-9b6a5073861a\") " Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.326510 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4" (OuterVolumeSpecName: "kube-api-access-gqjm4") pod "44b73a04-c864-44ed-b88d-9b6a5073861a" (UID: "44b73a04-c864-44ed-b88d-9b6a5073861a"). InnerVolumeSpecName "kube-api-access-gqjm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.330121 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg" (OuterVolumeSpecName: "kube-api-access-grdmg") pod "e0a447bc-40d5-401c-80c8-14d5d5385805" (UID: "e0a447bc-40d5-401c-80c8-14d5d5385805"). InnerVolumeSpecName "kube-api-access-grdmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.330202 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts" (OuterVolumeSpecName: "scripts") pod "44b73a04-c864-44ed-b88d-9b6a5073861a" (UID: "44b73a04-c864-44ed-b88d-9b6a5073861a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.352857 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data" (OuterVolumeSpecName: "config-data") pod "e0a447bc-40d5-401c-80c8-14d5d5385805" (UID: "e0a447bc-40d5-401c-80c8-14d5d5385805"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.357500 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data" (OuterVolumeSpecName: "config-data") pod "44b73a04-c864-44ed-b88d-9b6a5073861a" (UID: "44b73a04-c864-44ed-b88d-9b6a5073861a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.361400 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0a447bc-40d5-401c-80c8-14d5d5385805" (UID: "e0a447bc-40d5-401c-80c8-14d5d5385805"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.372527 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44b73a04-c864-44ed-b88d-9b6a5073861a" (UID: "44b73a04-c864-44ed-b88d-9b6a5073861a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424200 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424257 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424273 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44b73a04-c864-44ed-b88d-9b6a5073861a-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424288 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grdmg\" (UniqueName: \"kubernetes.io/projected/e0a447bc-40d5-401c-80c8-14d5d5385805-kube-api-access-grdmg\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424303 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqjm4\" (UniqueName: \"kubernetes.io/projected/44b73a04-c864-44ed-b88d-9b6a5073861a-kube-api-access-gqjm4\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424314 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.424326 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a447bc-40d5-401c-80c8-14d5d5385805-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:30 crc kubenswrapper[4867]: W1212 07:10:30.694353 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4210192a_fbe4_4a91_981e_112ecf49a237.slice/crio-1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8 WatchSource:0}: Error finding container 1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8: Status 404 returned error can't find the container with id 1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8 Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.697014 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.710607 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerStarted","Data":"1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8"} Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.713046 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e0a447bc-40d5-401c-80c8-14d5d5385805","Type":"ContainerDied","Data":"0ca5b74d791cb2896066fe7d370394b5a862d406544245447f4dcfc041d7605d"} Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.713067 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.713087 4867 scope.go:117] "RemoveContainer" containerID="6b73343f07d8e0ffe2561710edb199fe0ff3942f477bcbaa5d42a9ebf85bc561" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.721272 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" event={"ID":"44b73a04-c864-44ed-b88d-9b6a5073861a","Type":"ContainerDied","Data":"a7a5b947e5e8bdac366d6c7683882c037750719bd78663be978163aae7f9e718"} Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.721310 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7a5b947e5e8bdac366d6c7683882c037750719bd78663be978163aae7f9e718" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.721363 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dlfmg" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.768606 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.790899 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.800190 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: E1212 07:10:30.800761 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b73a04-c864-44ed-b88d-9b6a5073861a" containerName="nova-cell1-conductor-db-sync" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.800807 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b73a04-c864-44ed-b88d-9b6a5073861a" containerName="nova-cell1-conductor-db-sync" Dec 12 07:10:30 crc kubenswrapper[4867]: E1212 07:10:30.800821 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a447bc-40d5-401c-80c8-14d5d5385805" containerName="nova-scheduler-scheduler" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.800829 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a447bc-40d5-401c-80c8-14d5d5385805" containerName="nova-scheduler-scheduler" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.801066 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a447bc-40d5-401c-80c8-14d5d5385805" containerName="nova-scheduler-scheduler" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.801096 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b73a04-c864-44ed-b88d-9b6a5073861a" containerName="nova-cell1-conductor-db-sync" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.801952 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.810108 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.819388 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.830750 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.831958 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.835030 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.857252 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a447bc-40d5-401c-80c8-14d5d5385805" path="/var/lib/kubelet/pods/e0a447bc-40d5-401c-80c8-14d5d5385805/volumes" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.858871 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d334ec-a346-4490-80b3-9a88bc2b91e7" path="/var/lib/kubelet/pods/e1d334ec-a346-4490-80b3-9a88bc2b91e7/volumes" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.859497 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.934879 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.935005 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.935067 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh8rg\" (UniqueName: \"kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.935122 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.935155 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rjs6\" (UniqueName: \"kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:30 crc kubenswrapper[4867]: I1212 07:10:30.935175 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.037313 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rjs6\" (UniqueName: \"kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.037521 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.037669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.037888 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.038082 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh8rg\" (UniqueName: \"kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.038702 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.042921 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.044851 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.045435 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.045824 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.075674 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh8rg\" (UniqueName: \"kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg\") pod \"nova-scheduler-0\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.076108 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rjs6\" (UniqueName: \"kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6\") pod \"nova-cell1-conductor-0\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.191981 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.206494 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.669576 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:10:31 crc kubenswrapper[4867]: W1212 07:10:31.674772 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod366342dd_c366_4639_9f6d_6444f38fed68.slice/crio-abf715b11364c9206fbaf205c1a4c22aae12ace87b8d0a6d051596c0030a34c1 WatchSource:0}: Error finding container abf715b11364c9206fbaf205c1a4c22aae12ace87b8d0a6d051596c0030a34c1: Status 404 returned error can't find the container with id abf715b11364c9206fbaf205c1a4c22aae12ace87b8d0a6d051596c0030a34c1 Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.735060 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"366342dd-c366-4639-9f6d-6444f38fed68","Type":"ContainerStarted","Data":"abf715b11364c9206fbaf205c1a4c22aae12ace87b8d0a6d051596c0030a34c1"} Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.736943 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerStarted","Data":"eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f"} Dec 12 07:10:31 crc kubenswrapper[4867]: I1212 07:10:31.790381 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:10:31 crc kubenswrapper[4867]: W1212 07:10:31.791359 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09bd736a_175d_4a34_b0eb_714e10bd2207.slice/crio-09b19851156d33b364362c5216d737cf1bf604f291c379f0b7b8af6db1a9b9a1 WatchSource:0}: Error finding container 09b19851156d33b364362c5216d737cf1bf604f291c379f0b7b8af6db1a9b9a1: Status 404 returned error can't find the container with id 09b19851156d33b364362c5216d737cf1bf604f291c379f0b7b8af6db1a9b9a1 Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.754894 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerID="f2d06888e1824b50ccb88e51caad856d195a7d84b98146b6a12ccf98f22cdf19" exitCode=0 Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.755552 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerDied","Data":"f2d06888e1824b50ccb88e51caad856d195a7d84b98146b6a12ccf98f22cdf19"} Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.758308 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"366342dd-c366-4639-9f6d-6444f38fed68","Type":"ContainerStarted","Data":"00a292211a141c686e94c55059b96c051c4cc645dd2e169e1b511c3b048cd336"} Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.761663 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.768987 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"09bd736a-175d-4a34-b0eb-714e10bd2207","Type":"ContainerStarted","Data":"6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194"} Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.769048 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"09bd736a-175d-4a34-b0eb-714e10bd2207","Type":"ContainerStarted","Data":"09b19851156d33b364362c5216d737cf1bf604f291c379f0b7b8af6db1a9b9a1"} Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.772488 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerStarted","Data":"1671dbf245b4faa8dcbc3ca3be38cf020c46644501560453930a7a1ee4255c29"} Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.786168 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.786141497 podStartE2EDuration="2.786141497s" podCreationTimestamp="2025-12-12 07:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:32.781017071 +0000 UTC m=+1320.352398340" watchObservedRunningTime="2025-12-12 07:10:32.786141497 +0000 UTC m=+1320.357522766" Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.817596 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.817568439 podStartE2EDuration="3.817568439s" podCreationTimestamp="2025-12-12 07:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:32.810347309 +0000 UTC m=+1320.381728588" watchObservedRunningTime="2025-12-12 07:10:32.817568439 +0000 UTC m=+1320.388949708" Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.842202 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.84217619 podStartE2EDuration="2.84217619s" podCreationTimestamp="2025-12-12 07:10:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:32.838474218 +0000 UTC m=+1320.409855487" watchObservedRunningTime="2025-12-12 07:10:32.84217619 +0000 UTC m=+1320.413557459" Dec 12 07:10:32 crc kubenswrapper[4867]: I1212 07:10:32.956956 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.080101 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle\") pod \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.080548 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data\") pod \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.080626 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs\") pod \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.080670 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5948\" (UniqueName: \"kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948\") pod \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\" (UID: \"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8\") " Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.081022 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs" (OuterVolumeSpecName: "logs") pod "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" (UID: "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.081627 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.086691 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948" (OuterVolumeSpecName: "kube-api-access-b5948") pod "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" (UID: "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8"). InnerVolumeSpecName "kube-api-access-b5948". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.109831 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" (UID: "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.122642 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data" (OuterVolumeSpecName: "config-data") pod "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" (UID: "ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.183009 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.183053 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5948\" (UniqueName: \"kubernetes.io/projected/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-kube-api-access-b5948\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.183068 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.785984 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8","Type":"ContainerDied","Data":"46a3ea400c92dfe1e5ea9101fe63b45fd9f0a1ea2b3fef8c230395f3f20eba4b"} Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.786480 4867 scope.go:117] "RemoveContainer" containerID="f2d06888e1824b50ccb88e51caad856d195a7d84b98146b6a12ccf98f22cdf19" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.786694 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.825377 4867 scope.go:117] "RemoveContainer" containerID="d7fd8d67fa6de3d31075f82dfc86b3e76c705c975db03a4058b3357edc6f3b2a" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.829378 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.873669 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.918457 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:33 crc kubenswrapper[4867]: E1212 07:10:33.918948 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-api" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.918962 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-api" Dec 12 07:10:33 crc kubenswrapper[4867]: E1212 07:10:33.918998 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-log" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.919006 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-log" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.919194 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-api" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.919232 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" containerName="nova-api-log" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.920329 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.923660 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 07:10:33 crc kubenswrapper[4867]: I1212 07:10:33.949905 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.006573 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zkrq\" (UniqueName: \"kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.006621 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.006712 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.006743 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.108370 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zkrq\" (UniqueName: \"kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.108425 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.108471 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.108489 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.108978 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.113646 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.122788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.125588 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zkrq\" (UniqueName: \"kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq\") pod \"nova-api-0\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.253677 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.829765 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.859199 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8" path="/var/lib/kubelet/pods/ce7d65fe-c396-4888-b7cc-fb96fbcbc3e8/volumes" Dec 12 07:10:34 crc kubenswrapper[4867]: I1212 07:10:34.863074 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.166484 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.166597 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.806169 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerStarted","Data":"ad13beb9b0eb6c08814e4cf600123caebddf519df9b9d8e4befbebd8a00b8b7e"} Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.806570 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerStarted","Data":"a259f11297f2dfd7d146b05a77f93ab221c04b86877370a7964620c5892daa6d"} Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.806587 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerStarted","Data":"b0a8b2ab461805e0d76ca17e560466a85bad63fd25755969cd4467d30edd2ece"} Dec 12 07:10:35 crc kubenswrapper[4867]: I1212 07:10:35.828684 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.828661362 podStartE2EDuration="2.828661362s" podCreationTimestamp="2025-12-12 07:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:35.822973161 +0000 UTC m=+1323.394354450" watchObservedRunningTime="2025-12-12 07:10:35.828661362 +0000 UTC m=+1323.400042631" Dec 12 07:10:36 crc kubenswrapper[4867]: I1212 07:10:36.192964 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 07:10:38 crc kubenswrapper[4867]: I1212 07:10:38.582711 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:38 crc kubenswrapper[4867]: I1212 07:10:38.583216 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="344aed5c-2f3f-405c-a961-c31e19468100" containerName="kube-state-metrics" containerID="cri-o://f31de2a58bb8c0627835f82641e74392d8eaa399cc5779ec9d25f0762802a2c6" gracePeriod=30 Dec 12 07:10:38 crc kubenswrapper[4867]: I1212 07:10:38.839499 4867 generic.go:334] "Generic (PLEG): container finished" podID="344aed5c-2f3f-405c-a961-c31e19468100" containerID="f31de2a58bb8c0627835f82641e74392d8eaa399cc5779ec9d25f0762802a2c6" exitCode=2 Dec 12 07:10:38 crc kubenswrapper[4867]: I1212 07:10:38.851133 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"344aed5c-2f3f-405c-a961-c31e19468100","Type":"ContainerDied","Data":"f31de2a58bb8c0627835f82641e74392d8eaa399cc5779ec9d25f0762802a2c6"} Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.082362 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.245129 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnk2g\" (UniqueName: \"kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g\") pod \"344aed5c-2f3f-405c-a961-c31e19468100\" (UID: \"344aed5c-2f3f-405c-a961-c31e19468100\") " Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.250913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g" (OuterVolumeSpecName: "kube-api-access-tnk2g") pod "344aed5c-2f3f-405c-a961-c31e19468100" (UID: "344aed5c-2f3f-405c-a961-c31e19468100"). InnerVolumeSpecName "kube-api-access-tnk2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.348027 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnk2g\" (UniqueName: \"kubernetes.io/projected/344aed5c-2f3f-405c-a961-c31e19468100-kube-api-access-tnk2g\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.850403 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"344aed5c-2f3f-405c-a961-c31e19468100","Type":"ContainerDied","Data":"3e166c4ab0b1a8eec98fe03e56315537b792052fef2ad7865a92f1bd078dc22e"} Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.851489 4867 scope.go:117] "RemoveContainer" containerID="f31de2a58bb8c0627835f82641e74392d8eaa399cc5779ec9d25f0762802a2c6" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.850440 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.906372 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.917889 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.925660 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:39 crc kubenswrapper[4867]: E1212 07:10:39.926157 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344aed5c-2f3f-405c-a961-c31e19468100" containerName="kube-state-metrics" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.926184 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="344aed5c-2f3f-405c-a961-c31e19468100" containerName="kube-state-metrics" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.926442 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="344aed5c-2f3f-405c-a961-c31e19468100" containerName="kube-state-metrics" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.927421 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.930585 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.930622 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 12 07:10:39 crc kubenswrapper[4867]: I1212 07:10:39.942023 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.060434 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.060478 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cjs2\" (UniqueName: \"kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.060578 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.060908 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.162787 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.162906 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.162954 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.162997 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cjs2\" (UniqueName: \"kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.167540 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.167603 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.171335 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.173938 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.173964 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.181461 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cjs2\" (UniqueName: \"kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2\") pod \"kube-state-metrics-0\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.251466 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.402035 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.405547 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-central-agent" containerID="cri-o://0e9ae4f70c87c4971daedb51b9b5e5e54d2785732a42907c01159ee86492c662" gracePeriod=30 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.405657 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="sg-core" containerID="cri-o://6328805576d443af9afecb5b9296d72aa88d48e9f0c4504a04073e33616ff1fa" gracePeriod=30 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.405770 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="proxy-httpd" containerID="cri-o://edac33da98bf50816eab183b68a6de9858d9eaa542a52c8b09931db9f11a16d1" gracePeriod=30 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.405900 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-notification-agent" containerID="cri-o://9362ae3393372fcca9fb0c7bb91660a1a6e4882a5e215ffb2b52e9f6c1bff4cb" gracePeriod=30 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.851439 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="344aed5c-2f3f-405c-a961-c31e19468100" path="/var/lib/kubelet/pods/344aed5c-2f3f-405c-a961-c31e19468100/volumes" Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.884621 4867 generic.go:334] "Generic (PLEG): container finished" podID="442a5a33-bea0-4751-b675-92d03e288712" containerID="edac33da98bf50816eab183b68a6de9858d9eaa542a52c8b09931db9f11a16d1" exitCode=0 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.884670 4867 generic.go:334] "Generic (PLEG): container finished" podID="442a5a33-bea0-4751-b675-92d03e288712" containerID="6328805576d443af9afecb5b9296d72aa88d48e9f0c4504a04073e33616ff1fa" exitCode=2 Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.884760 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerDied","Data":"edac33da98bf50816eab183b68a6de9858d9eaa542a52c8b09931db9f11a16d1"} Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.884797 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerDied","Data":"6328805576d443af9afecb5b9296d72aa88d48e9f0c4504a04073e33616ff1fa"} Dec 12 07:10:40 crc kubenswrapper[4867]: I1212 07:10:40.923244 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.181509 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.182038 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.192199 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.228811 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.236847 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.903525 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"437a0623-5c6b-4a29-8da6-f23381fce0ec","Type":"ContainerStarted","Data":"a697abf1dfb4a8d1fd01effec382935f85b0f50a5309476015d25b7269a5f56e"} Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.904165 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"437a0623-5c6b-4a29-8da6-f23381fce0ec","Type":"ContainerStarted","Data":"d03b8a8f951c2b034df4288e68257eee6aeac9891b694168ecfaded3896fa122"} Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.904209 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.909159 4867 generic.go:334] "Generic (PLEG): container finished" podID="442a5a33-bea0-4751-b675-92d03e288712" containerID="0e9ae4f70c87c4971daedb51b9b5e5e54d2785732a42907c01159ee86492c662" exitCode=0 Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.909285 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerDied","Data":"0e9ae4f70c87c4971daedb51b9b5e5e54d2785732a42907c01159ee86492c662"} Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.920762 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.5602154649999997 podStartE2EDuration="2.920733493s" podCreationTimestamp="2025-12-12 07:10:39 +0000 UTC" firstStartedPulling="2025-12-12 07:10:40.92643751 +0000 UTC m=+1328.497818779" lastFinishedPulling="2025-12-12 07:10:41.286955538 +0000 UTC m=+1328.858336807" observedRunningTime="2025-12-12 07:10:41.920413036 +0000 UTC m=+1329.491794325" watchObservedRunningTime="2025-12-12 07:10:41.920733493 +0000 UTC m=+1329.492114762" Dec 12 07:10:41 crc kubenswrapper[4867]: I1212 07:10:41.957056 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 07:10:44 crc kubenswrapper[4867]: I1212 07:10:44.254972 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:10:44 crc kubenswrapper[4867]: I1212 07:10:44.255310 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:10:44 crc kubenswrapper[4867]: I1212 07:10:44.951007 4867 generic.go:334] "Generic (PLEG): container finished" podID="442a5a33-bea0-4751-b675-92d03e288712" containerID="9362ae3393372fcca9fb0c7bb91660a1a6e4882a5e215ffb2b52e9f6c1bff4cb" exitCode=0 Dec 12 07:10:44 crc kubenswrapper[4867]: I1212 07:10:44.951087 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerDied","Data":"9362ae3393372fcca9fb0c7bb91660a1a6e4882a5e215ffb2b52e9f6c1bff4cb"} Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.218801 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.338551 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.338579 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385458 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385519 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385547 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385613 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385668 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkd6d\" (UniqueName: \"kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385733 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.385790 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd\") pod \"442a5a33-bea0-4751-b675-92d03e288712\" (UID: \"442a5a33-bea0-4751-b675-92d03e288712\") " Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.386241 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.386654 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.407931 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d" (OuterVolumeSpecName: "kube-api-access-mkd6d") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "kube-api-access-mkd6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.408083 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts" (OuterVolumeSpecName: "scripts") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.420663 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.476315 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488123 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488174 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488189 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488202 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkd6d\" (UniqueName: \"kubernetes.io/projected/442a5a33-bea0-4751-b675-92d03e288712-kube-api-access-mkd6d\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488214 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.488247 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/442a5a33-bea0-4751-b675-92d03e288712-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.519246 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data" (OuterVolumeSpecName: "config-data") pod "442a5a33-bea0-4751-b675-92d03e288712" (UID: "442a5a33-bea0-4751-b675-92d03e288712"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.590322 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/442a5a33-bea0-4751-b675-92d03e288712-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.966525 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"442a5a33-bea0-4751-b675-92d03e288712","Type":"ContainerDied","Data":"f98297e7f805ae8effb2c1dab9f7b5595a042caab60f480043966d9642293c30"} Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.966587 4867 scope.go:117] "RemoveContainer" containerID="edac33da98bf50816eab183b68a6de9858d9eaa542a52c8b09931db9f11a16d1" Dec 12 07:10:45 crc kubenswrapper[4867]: I1212 07:10:45.966735 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.010422 4867 scope.go:117] "RemoveContainer" containerID="6328805576d443af9afecb5b9296d72aa88d48e9f0c4504a04073e33616ff1fa" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.010806 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.023325 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.038932 4867 scope.go:117] "RemoveContainer" containerID="9362ae3393372fcca9fb0c7bb91660a1a6e4882a5e215ffb2b52e9f6c1bff4cb" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.040708 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:46 crc kubenswrapper[4867]: E1212 07:10:46.041148 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="proxy-httpd" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041166 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="proxy-httpd" Dec 12 07:10:46 crc kubenswrapper[4867]: E1212 07:10:46.041180 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="sg-core" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041187 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="sg-core" Dec 12 07:10:46 crc kubenswrapper[4867]: E1212 07:10:46.041200 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-notification-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041206 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-notification-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: E1212 07:10:46.041266 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-central-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041273 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-central-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041443 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="proxy-httpd" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041458 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="sg-core" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041470 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-notification-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.041486 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="442a5a33-bea0-4751-b675-92d03e288712" containerName="ceilometer-central-agent" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.043089 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.053046 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.053723 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.053930 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.066425 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.116203 4867 scope.go:117] "RemoveContainer" containerID="0e9ae4f70c87c4971daedb51b9b5e5e54d2785732a42907c01159ee86492c662" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.207957 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208107 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208137 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208197 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ztsr\" (UniqueName: \"kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208320 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208347 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208414 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.208435 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310478 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310533 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310580 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ztsr\" (UniqueName: \"kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310607 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310628 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310683 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310703 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.310764 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.311347 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.311955 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.314320 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.315615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.315935 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.316288 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.325593 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.326274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ztsr\" (UniqueName: \"kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr\") pod \"ceilometer-0\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.421244 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.855003 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442a5a33-bea0-4751-b675-92d03e288712" path="/var/lib/kubelet/pods/442a5a33-bea0-4751-b675-92d03e288712/volumes" Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.892944 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:46 crc kubenswrapper[4867]: I1212 07:10:46.977443 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerStarted","Data":"7ad91e513231a6c2c7cbf7c11d27e5cc791d5c29d189b830e28efa7b31172450"} Dec 12 07:10:48 crc kubenswrapper[4867]: I1212 07:10:48.998988 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerStarted","Data":"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a"} Dec 12 07:10:50 crc kubenswrapper[4867]: I1212 07:10:50.171327 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 07:10:50 crc kubenswrapper[4867]: I1212 07:10:50.173680 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 07:10:50 crc kubenswrapper[4867]: I1212 07:10:50.176692 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 07:10:50 crc kubenswrapper[4867]: I1212 07:10:50.271149 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 12 07:10:51 crc kubenswrapper[4867]: I1212 07:10:51.025722 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerStarted","Data":"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5"} Dec 12 07:10:51 crc kubenswrapper[4867]: I1212 07:10:51.036085 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 07:10:52 crc kubenswrapper[4867]: I1212 07:10:52.037121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerStarted","Data":"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a"} Dec 12 07:10:52 crc kubenswrapper[4867]: I1212 07:10:52.990249 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.048627 4867 generic.go:334] "Generic (PLEG): container finished" podID="fd0ec4ba-7992-4650-a749-5fd94190938d" containerID="d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8" exitCode=137 Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.049612 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.050113 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd0ec4ba-7992-4650-a749-5fd94190938d","Type":"ContainerDied","Data":"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8"} Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.050153 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fd0ec4ba-7992-4650-a749-5fd94190938d","Type":"ContainerDied","Data":"aa590cc23fc2b3a641f03ed11798ec7488deb7992789614faffd7122cbf49b1c"} Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.050173 4867 scope.go:117] "RemoveContainer" containerID="d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.075417 4867 scope.go:117] "RemoveContainer" containerID="d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8" Dec 12 07:10:53 crc kubenswrapper[4867]: E1212 07:10:53.075952 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8\": container with ID starting with d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8 not found: ID does not exist" containerID="d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.076019 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8"} err="failed to get container status \"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8\": rpc error: code = NotFound desc = could not find container \"d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8\": container with ID starting with d47aadef5217b2cb4705ecec835be343da27bd5d2c18f5b5675e5faa0f028ff8 not found: ID does not exist" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.170346 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j9g2\" (UniqueName: \"kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2\") pod \"fd0ec4ba-7992-4650-a749-5fd94190938d\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.170611 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle\") pod \"fd0ec4ba-7992-4650-a749-5fd94190938d\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.170698 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data\") pod \"fd0ec4ba-7992-4650-a749-5fd94190938d\" (UID: \"fd0ec4ba-7992-4650-a749-5fd94190938d\") " Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.175908 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2" (OuterVolumeSpecName: "kube-api-access-5j9g2") pod "fd0ec4ba-7992-4650-a749-5fd94190938d" (UID: "fd0ec4ba-7992-4650-a749-5fd94190938d"). InnerVolumeSpecName "kube-api-access-5j9g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.205108 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data" (OuterVolumeSpecName: "config-data") pod "fd0ec4ba-7992-4650-a749-5fd94190938d" (UID: "fd0ec4ba-7992-4650-a749-5fd94190938d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.220323 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd0ec4ba-7992-4650-a749-5fd94190938d" (UID: "fd0ec4ba-7992-4650-a749-5fd94190938d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.273333 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.273364 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0ec4ba-7992-4650-a749-5fd94190938d-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.273387 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j9g2\" (UniqueName: \"kubernetes.io/projected/fd0ec4ba-7992-4650-a749-5fd94190938d-kube-api-access-5j9g2\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.389015 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.398589 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.409394 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:53 crc kubenswrapper[4867]: E1212 07:10:53.410088 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd0ec4ba-7992-4650-a749-5fd94190938d" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.410111 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd0ec4ba-7992-4650-a749-5fd94190938d" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.410395 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd0ec4ba-7992-4650-a749-5fd94190938d" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.411336 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.413950 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.414170 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.416453 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.445783 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.578410 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.578758 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.578780 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.578816 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.579465 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xl4x\" (UniqueName: \"kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.681396 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.681495 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xl4x\" (UniqueName: \"kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.681584 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.681614 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.681630 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.686615 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.686798 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.686963 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.687587 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.699913 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xl4x\" (UniqueName: \"kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:53 crc kubenswrapper[4867]: I1212 07:10:53.736888 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.060717 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerStarted","Data":"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267"} Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.061180 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.119439 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.573603853 podStartE2EDuration="8.119399897s" podCreationTimestamp="2025-12-12 07:10:46 +0000 UTC" firstStartedPulling="2025-12-12 07:10:46.897828493 +0000 UTC m=+1334.469209762" lastFinishedPulling="2025-12-12 07:10:53.443624537 +0000 UTC m=+1341.015005806" observedRunningTime="2025-12-12 07:10:54.085088874 +0000 UTC m=+1341.656470143" watchObservedRunningTime="2025-12-12 07:10:54.119399897 +0000 UTC m=+1341.690781186" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.182664 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.263839 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.265137 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.271954 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.289817 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 07:10:54 crc kubenswrapper[4867]: I1212 07:10:54.851650 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd0ec4ba-7992-4650-a749-5fd94190938d" path="/var/lib/kubelet/pods/fd0ec4ba-7992-4650-a749-5fd94190938d/volumes" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.076928 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf84c603-ce1e-42be-a595-b27ddc77880c","Type":"ContainerStarted","Data":"327346d748281830108225a6a9e113b8722507f57f85ac4bbb3af491d43078a5"} Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.078299 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf84c603-ce1e-42be-a595-b27ddc77880c","Type":"ContainerStarted","Data":"fe6d5d260958339042003dfb25398c1fed02916feffc9dd34e55df3164025009"} Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.078439 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.085547 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.099453 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.099436707 podStartE2EDuration="2.099436707s" podCreationTimestamp="2025-12-12 07:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:55.097502048 +0000 UTC m=+1342.668883347" watchObservedRunningTime="2025-12-12 07:10:55.099436707 +0000 UTC m=+1342.670817976" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.292368 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.294444 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.310814 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.418871 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.419003 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.419053 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.419096 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.419202 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br7h8\" (UniqueName: \"kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.419456 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.521837 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.521909 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.521932 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.521970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br7h8\" (UniqueName: \"kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.522006 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.522103 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.522718 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.522954 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.523396 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.523633 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.524273 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.549676 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br7h8\" (UniqueName: \"kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8\") pod \"dnsmasq-dns-fcd6f8f8f-whlhl\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:55 crc kubenswrapper[4867]: I1212 07:10:55.628188 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:56 crc kubenswrapper[4867]: I1212 07:10:56.145416 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.093373 4867 generic.go:334] "Generic (PLEG): container finished" podID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerID="334aa3a251311e5ad85b964132678178f0417b43878df9f044e9cc662fc15fbd" exitCode=0 Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.093424 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" event={"ID":"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8","Type":"ContainerDied","Data":"334aa3a251311e5ad85b964132678178f0417b43878df9f044e9cc662fc15fbd"} Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.093702 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" event={"ID":"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8","Type":"ContainerStarted","Data":"5bd17181a1f2fd4f740b3706415ba0cb8648d14a66fe5f0309d85a79d3e28d31"} Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.619564 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.620124 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-central-agent" containerID="cri-o://2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" gracePeriod=30 Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.620146 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="proxy-httpd" containerID="cri-o://3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" gracePeriod=30 Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.620200 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-notification-agent" containerID="cri-o://1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" gracePeriod=30 Dec 12 07:10:57 crc kubenswrapper[4867]: I1212 07:10:57.620216 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="sg-core" containerID="cri-o://62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" gracePeriod=30 Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.106514 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" event={"ID":"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8","Type":"ContainerStarted","Data":"5b998256ed7e715606082bd623b651120a069bd1bdab372dad3314ef077aeaaf"} Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.107043 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.109636 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerID="3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" exitCode=0 Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.109663 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerID="62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" exitCode=2 Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.109680 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerDied","Data":"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267"} Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.109696 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerDied","Data":"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a"} Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.131026 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" podStartSLOduration=3.131003998 podStartE2EDuration="3.131003998s" podCreationTimestamp="2025-12-12 07:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:10:58.125777768 +0000 UTC m=+1345.697159047" watchObservedRunningTime="2025-12-12 07:10:58.131003998 +0000 UTC m=+1345.702385267" Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.737340 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.829344 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.830055 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-log" containerID="cri-o://a259f11297f2dfd7d146b05a77f93ab221c04b86877370a7964620c5892daa6d" gracePeriod=30 Dec 12 07:10:58 crc kubenswrapper[4867]: I1212 07:10:58.830531 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-api" containerID="cri-o://ad13beb9b0eb6c08814e4cf600123caebddf519df9b9d8e4befbebd8a00b8b7e" gracePeriod=30 Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.009654 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.092981 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093051 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093077 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093167 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093189 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093305 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093335 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ztsr\" (UniqueName: \"kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr\") pod \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\" (UID: \"c2e52dd5-e5bb-4fa2-b540-4ae70830a265\") " Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093820 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.093856 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.099519 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr" (OuterVolumeSpecName: "kube-api-access-5ztsr") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "kube-api-access-5ztsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.108854 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts" (OuterVolumeSpecName: "scripts") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121646 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerID="1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" exitCode=0 Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121679 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerID="2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" exitCode=0 Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121716 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerDied","Data":"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5"} Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121742 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerDied","Data":"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a"} Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121752 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2e52dd5-e5bb-4fa2-b540-4ae70830a265","Type":"ContainerDied","Data":"7ad91e513231a6c2c7cbf7c11d27e5cc791d5c29d189b830e28efa7b31172450"} Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121767 4867 scope.go:117] "RemoveContainer" containerID="3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.121885 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.133560 4867 generic.go:334] "Generic (PLEG): container finished" podID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerID="a259f11297f2dfd7d146b05a77f93ab221c04b86877370a7964620c5892daa6d" exitCode=143 Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.133663 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerDied","Data":"a259f11297f2dfd7d146b05a77f93ab221c04b86877370a7964620c5892daa6d"} Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.135789 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.162363 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195375 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195414 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195428 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195438 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195450 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.195458 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ztsr\" (UniqueName: \"kubernetes.io/projected/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-kube-api-access-5ztsr\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.212672 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.225109 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data" (OuterVolumeSpecName: "config-data") pod "c2e52dd5-e5bb-4fa2-b540-4ae70830a265" (UID: "c2e52dd5-e5bb-4fa2-b540-4ae70830a265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.297615 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.297810 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2e52dd5-e5bb-4fa2-b540-4ae70830a265-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.317787 4867 scope.go:117] "RemoveContainer" containerID="62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.340935 4867 scope.go:117] "RemoveContainer" containerID="1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.367818 4867 scope.go:117] "RemoveContainer" containerID="2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.394088 4867 scope.go:117] "RemoveContainer" containerID="3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.394553 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267\": container with ID starting with 3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267 not found: ID does not exist" containerID="3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.394582 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267"} err="failed to get container status \"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267\": rpc error: code = NotFound desc = could not find container \"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267\": container with ID starting with 3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267 not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.394605 4867 scope.go:117] "RemoveContainer" containerID="62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.394921 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a\": container with ID starting with 62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a not found: ID does not exist" containerID="62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.394944 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a"} err="failed to get container status \"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a\": rpc error: code = NotFound desc = could not find container \"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a\": container with ID starting with 62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.394958 4867 scope.go:117] "RemoveContainer" containerID="1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.395141 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5\": container with ID starting with 1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5 not found: ID does not exist" containerID="1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395161 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5"} err="failed to get container status \"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5\": rpc error: code = NotFound desc = could not find container \"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5\": container with ID starting with 1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5 not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395174 4867 scope.go:117] "RemoveContainer" containerID="2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.395396 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a\": container with ID starting with 2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a not found: ID does not exist" containerID="2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395418 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a"} err="failed to get container status \"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a\": rpc error: code = NotFound desc = could not find container \"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a\": container with ID starting with 2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395431 4867 scope.go:117] "RemoveContainer" containerID="3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395657 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267"} err="failed to get container status \"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267\": rpc error: code = NotFound desc = could not find container \"3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267\": container with ID starting with 3ba6b5a426555a475100ae59bc24dc8256ca6f1774a7b37ef874fd6534308267 not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395674 4867 scope.go:117] "RemoveContainer" containerID="62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395980 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a"} err="failed to get container status \"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a\": rpc error: code = NotFound desc = could not find container \"62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a\": container with ID starting with 62c0d70fee4fb72a7da1eb6961951fde6a41b3c27a8f6e9718e0611da00aa00a not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.395999 4867 scope.go:117] "RemoveContainer" containerID="1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.396240 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5"} err="failed to get container status \"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5\": rpc error: code = NotFound desc = could not find container \"1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5\": container with ID starting with 1f8d8933353df1482440b0902228cd9725fbce5c0150c167b333b0435fc5fcd5 not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.396258 4867 scope.go:117] "RemoveContainer" containerID="2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.396469 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a"} err="failed to get container status \"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a\": rpc error: code = NotFound desc = could not find container \"2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a\": container with ID starting with 2ae3f3ffb0ae66c6dd3cedacd2790bea0508fca90a0a62186d8e7be463bbb57a not found: ID does not exist" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.456443 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.467264 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.481614 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.482112 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-central-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482139 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-central-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.482167 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="proxy-httpd" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482175 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="proxy-httpd" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.482188 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="sg-core" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482196 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="sg-core" Dec 12 07:10:59 crc kubenswrapper[4867]: E1212 07:10:59.482211 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-notification-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482219 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-notification-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482464 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-central-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482491 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="proxy-httpd" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482505 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="sg-core" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.482524 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" containerName="ceilometer-notification-agent" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.484739 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.487350 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.487350 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.487350 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.496857 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.605899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dkh6\" (UniqueName: \"kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.605978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606001 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606048 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606129 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606155 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.606180 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.707818 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.707902 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.707928 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.707952 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708018 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dkh6\" (UniqueName: \"kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708071 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708087 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708603 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.708914 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.713069 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.713075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.713653 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.714265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.715212 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.725796 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dkh6\" (UniqueName: \"kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6\") pod \"ceilometer-0\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " pod="openstack/ceilometer-0" Dec 12 07:10:59 crc kubenswrapper[4867]: I1212 07:10:59.816311 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:11:00 crc kubenswrapper[4867]: W1212 07:11:00.270318 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3950b22_ca71_44c2_89e6_1209a53d0cd3.slice/crio-2ae63e7053d55c1cd068e85c905d24ac70b4bc8581d9650b5919da596ed78aa2 WatchSource:0}: Error finding container 2ae63e7053d55c1cd068e85c905d24ac70b4bc8581d9650b5919da596ed78aa2: Status 404 returned error can't find the container with id 2ae63e7053d55c1cd068e85c905d24ac70b4bc8581d9650b5919da596ed78aa2 Dec 12 07:11:00 crc kubenswrapper[4867]: I1212 07:11:00.271656 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:00 crc kubenswrapper[4867]: I1212 07:11:00.874280 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e52dd5-e5bb-4fa2-b540-4ae70830a265" path="/var/lib/kubelet/pods/c2e52dd5-e5bb-4fa2-b540-4ae70830a265/volumes" Dec 12 07:11:01 crc kubenswrapper[4867]: I1212 07:11:01.153456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerStarted","Data":"2ae63e7053d55c1cd068e85c905d24ac70b4bc8581d9650b5919da596ed78aa2"} Dec 12 07:11:01 crc kubenswrapper[4867]: I1212 07:11:01.363750 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.173720 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerStarted","Data":"715f3afdac23fd09e49196492e04351d6f822ad0a7b572801de345b914b2a1c1"} Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.186756 4867 generic.go:334] "Generic (PLEG): container finished" podID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerID="ad13beb9b0eb6c08814e4cf600123caebddf519df9b9d8e4befbebd8a00b8b7e" exitCode=0 Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.186812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerDied","Data":"ad13beb9b0eb6c08814e4cf600123caebddf519df9b9d8e4befbebd8a00b8b7e"} Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.480345 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.569263 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs\") pod \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.569532 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zkrq\" (UniqueName: \"kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq\") pod \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.569909 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data\") pod \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.569971 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle\") pod \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\" (UID: \"35edd3ce-975f-4e8b-94b9-a5a441804ec2\") " Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.570355 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs" (OuterVolumeSpecName: "logs") pod "35edd3ce-975f-4e8b-94b9-a5a441804ec2" (UID: "35edd3ce-975f-4e8b-94b9-a5a441804ec2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.570751 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35edd3ce-975f-4e8b-94b9-a5a441804ec2-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.582583 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq" (OuterVolumeSpecName: "kube-api-access-8zkrq") pod "35edd3ce-975f-4e8b-94b9-a5a441804ec2" (UID: "35edd3ce-975f-4e8b-94b9-a5a441804ec2"). InnerVolumeSpecName "kube-api-access-8zkrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.630393 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data" (OuterVolumeSpecName: "config-data") pod "35edd3ce-975f-4e8b-94b9-a5a441804ec2" (UID: "35edd3ce-975f-4e8b-94b9-a5a441804ec2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.665503 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35edd3ce-975f-4e8b-94b9-a5a441804ec2" (UID: "35edd3ce-975f-4e8b-94b9-a5a441804ec2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.672618 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.672659 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35edd3ce-975f-4e8b-94b9-a5a441804ec2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:02 crc kubenswrapper[4867]: I1212 07:11:02.672674 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zkrq\" (UniqueName: \"kubernetes.io/projected/35edd3ce-975f-4e8b-94b9-a5a441804ec2-kube-api-access-8zkrq\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.197333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35edd3ce-975f-4e8b-94b9-a5a441804ec2","Type":"ContainerDied","Data":"b0a8b2ab461805e0d76ca17e560466a85bad63fd25755969cd4467d30edd2ece"} Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.197663 4867 scope.go:117] "RemoveContainer" containerID="ad13beb9b0eb6c08814e4cf600123caebddf519df9b9d8e4befbebd8a00b8b7e" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.197715 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.203532 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerStarted","Data":"1cbd7ae07f31048ed723481fba0f7b7f22a511f8cafc77bbfecf93da165d05a8"} Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.226681 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.232881 4867 scope.go:117] "RemoveContainer" containerID="a259f11297f2dfd7d146b05a77f93ab221c04b86877370a7964620c5892daa6d" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.250460 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.261315 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:03 crc kubenswrapper[4867]: E1212 07:11:03.262283 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-log" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.262305 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-log" Dec 12 07:11:03 crc kubenswrapper[4867]: E1212 07:11:03.262357 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-api" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.262365 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-api" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.262540 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-log" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.262567 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" containerName="nova-api-api" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.263496 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.266817 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.266903 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.266949 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.281780 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387357 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387419 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387596 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387666 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bttnk\" (UniqueName: \"kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387764 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.387937 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489187 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bttnk\" (UniqueName: \"kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489303 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489603 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489641 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489674 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489716 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.489963 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.501197 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.502131 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.507284 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.507739 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.512953 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bttnk\" (UniqueName: \"kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk\") pod \"nova-api-0\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.579905 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.737152 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:11:03 crc kubenswrapper[4867]: I1212 07:11:03.777828 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.042841 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:04 crc kubenswrapper[4867]: W1212 07:11:04.044676 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a136e3a_f31d_44fb_9922_e73e7886592f.slice/crio-cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9 WatchSource:0}: Error finding container cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9: Status 404 returned error can't find the container with id cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9 Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.215276 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerStarted","Data":"cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9"} Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.217523 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerStarted","Data":"253e8f652d0053f45aa89e8099a1d53ff79e1d5dfbb7086e369f0fb29d2905ef"} Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.233720 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.449273 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9c5ps"] Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.450858 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.453866 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.454110 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.460957 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9c5ps"] Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.610330 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976ls\" (UniqueName: \"kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.610608 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.610725 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.610747 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.712481 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.712547 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.712631 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976ls\" (UniqueName: \"kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.712681 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.718456 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.718610 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.719812 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.730116 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976ls\" (UniqueName: \"kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls\") pod \"nova-cell1-cell-mapping-9c5ps\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.849874 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35edd3ce-975f-4e8b-94b9-a5a441804ec2" path="/var/lib/kubelet/pods/35edd3ce-975f-4e8b-94b9-a5a441804ec2/volumes" Dec 12 07:11:04 crc kubenswrapper[4867]: I1212 07:11:04.977402 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.232479 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerStarted","Data":"36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8"} Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.232792 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerStarted","Data":"a0bedba8914b8a45bc1c0ddce256133a717fdad864ae53b01a0eb5e13c53466b"} Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238204 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-central-agent" containerID="cri-o://715f3afdac23fd09e49196492e04351d6f822ad0a7b572801de345b914b2a1c1" gracePeriod=30 Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238436 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="proxy-httpd" containerID="cri-o://11d1470839b0af49726e251b1c18e90b839155db336892a680b70848ec7eceaf" gracePeriod=30 Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238485 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="sg-core" containerID="cri-o://253e8f652d0053f45aa89e8099a1d53ff79e1d5dfbb7086e369f0fb29d2905ef" gracePeriod=30 Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238528 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-notification-agent" containerID="cri-o://1cbd7ae07f31048ed723481fba0f7b7f22a511f8cafc77bbfecf93da165d05a8" gracePeriod=30 Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238582 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerStarted","Data":"11d1470839b0af49726e251b1c18e90b839155db336892a680b70848ec7eceaf"} Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.238634 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.275740 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.275705703 podStartE2EDuration="2.275705703s" podCreationTimestamp="2025-12-12 07:11:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:11:05.263296724 +0000 UTC m=+1352.834677993" watchObservedRunningTime="2025-12-12 07:11:05.275705703 +0000 UTC m=+1352.847086972" Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.294946 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.662624817 podStartE2EDuration="6.294923999s" podCreationTimestamp="2025-12-12 07:10:59 +0000 UTC" firstStartedPulling="2025-12-12 07:11:00.273375846 +0000 UTC m=+1347.844757115" lastFinishedPulling="2025-12-12 07:11:04.905675018 +0000 UTC m=+1352.477056297" observedRunningTime="2025-12-12 07:11:05.289104325 +0000 UTC m=+1352.860485594" watchObservedRunningTime="2025-12-12 07:11:05.294923999 +0000 UTC m=+1352.866305258" Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.459811 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9c5ps"] Dec 12 07:11:05 crc kubenswrapper[4867]: W1212 07:11:05.469601 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91f25ce3_2ba0_49ce_8842_f26edb941ad5.slice/crio-03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d WatchSource:0}: Error finding container 03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d: Status 404 returned error can't find the container with id 03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.630282 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.714634 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:11:05 crc kubenswrapper[4867]: I1212 07:11:05.714917 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="dnsmasq-dns" containerID="cri-o://c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c" gracePeriod=10 Dec 12 07:11:06 crc kubenswrapper[4867]: I1212 07:11:06.250399 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerID="253e8f652d0053f45aa89e8099a1d53ff79e1d5dfbb7086e369f0fb29d2905ef" exitCode=2 Dec 12 07:11:06 crc kubenswrapper[4867]: I1212 07:11:06.250456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerDied","Data":"253e8f652d0053f45aa89e8099a1d53ff79e1d5dfbb7086e369f0fb29d2905ef"} Dec 12 07:11:06 crc kubenswrapper[4867]: I1212 07:11:06.252023 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9c5ps" event={"ID":"91f25ce3-2ba0-49ce-8842-f26edb941ad5","Type":"ContainerStarted","Data":"03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d"} Dec 12 07:11:06 crc kubenswrapper[4867]: I1212 07:11:06.956747 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.065634 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptlg8\" (UniqueName: \"kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.065834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.065897 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.065938 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.066115 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.066216 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc\") pod \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\" (UID: \"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9\") " Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.074638 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8" (OuterVolumeSpecName: "kube-api-access-ptlg8") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "kube-api-access-ptlg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.119815 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.127498 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.129353 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.133010 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config" (OuterVolumeSpecName: "config") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.145991 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" (UID: "c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169343 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169385 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169395 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptlg8\" (UniqueName: \"kubernetes.io/projected/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-kube-api-access-ptlg8\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169405 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169415 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.169426 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.265957 4867 generic.go:334] "Generic (PLEG): container finished" podID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerID="c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c" exitCode=0 Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.266025 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.266022 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" event={"ID":"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9","Type":"ContainerDied","Data":"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c"} Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.266108 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-647df7b8c5-47dxv" event={"ID":"c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9","Type":"ContainerDied","Data":"757a4b1d9238758b4288cf09717a12ffc0850f78a43bda807275549fe26be2e7"} Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.266128 4867 scope.go:117] "RemoveContainer" containerID="c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.278826 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerID="1cbd7ae07f31048ed723481fba0f7b7f22a511f8cafc77bbfecf93da165d05a8" exitCode=0 Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.278886 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerDied","Data":"1cbd7ae07f31048ed723481fba0f7b7f22a511f8cafc77bbfecf93da165d05a8"} Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.282148 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9c5ps" event={"ID":"91f25ce3-2ba0-49ce-8842-f26edb941ad5","Type":"ContainerStarted","Data":"ed60bf6d99ceafbd9fedf17a5214f14b39f75b17d2af05f1022ba59e01bf9864"} Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.305271 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9c5ps" podStartSLOduration=3.305251797 podStartE2EDuration="3.305251797s" podCreationTimestamp="2025-12-12 07:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:11:07.301866274 +0000 UTC m=+1354.873247543" watchObservedRunningTime="2025-12-12 07:11:07.305251797 +0000 UTC m=+1354.876633066" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.332335 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.335706 4867 scope.go:117] "RemoveContainer" containerID="09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.341996 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-647df7b8c5-47dxv"] Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.365649 4867 scope.go:117] "RemoveContainer" containerID="c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c" Dec 12 07:11:07 crc kubenswrapper[4867]: E1212 07:11:07.366032 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c\": container with ID starting with c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c not found: ID does not exist" containerID="c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.366058 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c"} err="failed to get container status \"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c\": rpc error: code = NotFound desc = could not find container \"c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c\": container with ID starting with c0e614e733377557a68ce356300a39ce8946d8b509134858b3a315c7c2e2de0c not found: ID does not exist" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.366079 4867 scope.go:117] "RemoveContainer" containerID="09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58" Dec 12 07:11:07 crc kubenswrapper[4867]: E1212 07:11:07.366341 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58\": container with ID starting with 09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58 not found: ID does not exist" containerID="09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58" Dec 12 07:11:07 crc kubenswrapper[4867]: I1212 07:11:07.366360 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58"} err="failed to get container status \"09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58\": rpc error: code = NotFound desc = could not find container \"09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58\": container with ID starting with 09f295d517d58817ffe3694787fb6217ceb461e94647055eba57ea41a4407c58 not found: ID does not exist" Dec 12 07:11:08 crc kubenswrapper[4867]: I1212 07:11:08.305152 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerID="715f3afdac23fd09e49196492e04351d6f822ad0a7b572801de345b914b2a1c1" exitCode=0 Dec 12 07:11:08 crc kubenswrapper[4867]: I1212 07:11:08.305246 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerDied","Data":"715f3afdac23fd09e49196492e04351d6f822ad0a7b572801de345b914b2a1c1"} Dec 12 07:11:08 crc kubenswrapper[4867]: I1212 07:11:08.847851 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" path="/var/lib/kubelet/pods/c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9/volumes" Dec 12 07:11:11 crc kubenswrapper[4867]: I1212 07:11:11.331623 4867 generic.go:334] "Generic (PLEG): container finished" podID="91f25ce3-2ba0-49ce-8842-f26edb941ad5" containerID="ed60bf6d99ceafbd9fedf17a5214f14b39f75b17d2af05f1022ba59e01bf9864" exitCode=0 Dec 12 07:11:11 crc kubenswrapper[4867]: I1212 07:11:11.331723 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9c5ps" event={"ID":"91f25ce3-2ba0-49ce-8842-f26edb941ad5","Type":"ContainerDied","Data":"ed60bf6d99ceafbd9fedf17a5214f14b39f75b17d2af05f1022ba59e01bf9864"} Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.734797 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.883575 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts\") pod \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.883655 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle\") pod \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.883737 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976ls\" (UniqueName: \"kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls\") pod \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.883810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data\") pod \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\" (UID: \"91f25ce3-2ba0-49ce-8842-f26edb941ad5\") " Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.890416 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts" (OuterVolumeSpecName: "scripts") pod "91f25ce3-2ba0-49ce-8842-f26edb941ad5" (UID: "91f25ce3-2ba0-49ce-8842-f26edb941ad5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.895657 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls" (OuterVolumeSpecName: "kube-api-access-976ls") pod "91f25ce3-2ba0-49ce-8842-f26edb941ad5" (UID: "91f25ce3-2ba0-49ce-8842-f26edb941ad5"). InnerVolumeSpecName "kube-api-access-976ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.919784 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data" (OuterVolumeSpecName: "config-data") pod "91f25ce3-2ba0-49ce-8842-f26edb941ad5" (UID: "91f25ce3-2ba0-49ce-8842-f26edb941ad5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.925454 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91f25ce3-2ba0-49ce-8842-f26edb941ad5" (UID: "91f25ce3-2ba0-49ce-8842-f26edb941ad5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.986466 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976ls\" (UniqueName: \"kubernetes.io/projected/91f25ce3-2ba0-49ce-8842-f26edb941ad5-kube-api-access-976ls\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.986501 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.986512 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:12 crc kubenswrapper[4867]: I1212 07:11:12.986525 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f25ce3-2ba0-49ce-8842-f26edb941ad5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.354842 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9c5ps" event={"ID":"91f25ce3-2ba0-49ce-8842-f26edb941ad5","Type":"ContainerDied","Data":"03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d"} Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.354897 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03d21a37b675d989ffbdfa4a2187225cae4a9f122e6627c8eda030b7e3d72c5d" Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.354940 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9c5ps" Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.553284 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.554129 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-log" containerID="cri-o://a0bedba8914b8a45bc1c0ddce256133a717fdad864ae53b01a0eb5e13c53466b" gracePeriod=30 Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.554195 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-api" containerID="cri-o://36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8" gracePeriod=30 Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.566008 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.566505 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerName="nova-scheduler-scheduler" containerID="cri-o://6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" gracePeriod=30 Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.663423 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.663711 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" containerID="cri-o://eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f" gracePeriod=30 Dec 12 07:11:13 crc kubenswrapper[4867]: I1212 07:11:13.663874 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" containerID="cri-o://1671dbf245b4faa8dcbc3ca3be38cf020c46644501560453930a7a1ee4255c29" gracePeriod=30 Dec 12 07:11:13 crc kubenswrapper[4867]: E1212 07:11:13.995102 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4210192a_fbe4_4a91_981e_112ecf49a237.slice/crio-eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a136e3a_f31d_44fb_9922_e73e7886592f.slice/crio-conmon-36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4210192a_fbe4_4a91_981e_112ecf49a237.slice/crio-conmon-eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f.scope\": RecentStats: unable to find data in memory cache]" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.379968 4867 generic.go:334] "Generic (PLEG): container finished" podID="4210192a-fbe4-4a91-981e-112ecf49a237" containerID="eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f" exitCode=143 Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.380612 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerDied","Data":"eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f"} Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383458 4867 generic.go:334] "Generic (PLEG): container finished" podID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerID="36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8" exitCode=0 Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383482 4867 generic.go:334] "Generic (PLEG): container finished" podID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerID="a0bedba8914b8a45bc1c0ddce256133a717fdad864ae53b01a0eb5e13c53466b" exitCode=143 Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383499 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerDied","Data":"36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8"} Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383517 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerDied","Data":"a0bedba8914b8a45bc1c0ddce256133a717fdad864ae53b01a0eb5e13c53466b"} Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383526 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a136e3a-f31d-44fb-9922-e73e7886592f","Type":"ContainerDied","Data":"cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9"} Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.383537 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf52622f3c4461dcfc5a562f3106051f796367bbbfa47ce49bbafa56adcea5d9" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.415638 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.520856 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.520939 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.520977 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.521133 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.521186 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.521246 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bttnk\" (UniqueName: \"kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk\") pod \"0a136e3a-f31d-44fb-9922-e73e7886592f\" (UID: \"0a136e3a-f31d-44fb-9922-e73e7886592f\") " Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.522197 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs" (OuterVolumeSpecName: "logs") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.526614 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk" (OuterVolumeSpecName: "kube-api-access-bttnk") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "kube-api-access-bttnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.552301 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.559212 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data" (OuterVolumeSpecName: "config-data") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.579602 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.581337 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a136e3a-f31d-44fb-9922-e73e7886592f" (UID: "0a136e3a-f31d-44fb-9922-e73e7886592f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624211 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624260 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a136e3a-f31d-44fb-9922-e73e7886592f-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624273 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bttnk\" (UniqueName: \"kubernetes.io/projected/0a136e3a-f31d-44fb-9922-e73e7886592f-kube-api-access-bttnk\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624287 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624297 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:14 crc kubenswrapper[4867]: I1212 07:11:14.624307 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a136e3a-f31d-44fb-9922-e73e7886592f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.395774 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.441651 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.450816 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.473736 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:15 crc kubenswrapper[4867]: E1212 07:11:15.474595 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-log" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.474700 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-log" Dec 12 07:11:15 crc kubenswrapper[4867]: E1212 07:11:15.474776 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="init" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.474862 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="init" Dec 12 07:11:15 crc kubenswrapper[4867]: E1212 07:11:15.474958 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-api" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475037 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-api" Dec 12 07:11:15 crc kubenswrapper[4867]: E1212 07:11:15.475132 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f25ce3-2ba0-49ce-8842-f26edb941ad5" containerName="nova-manage" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475194 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f25ce3-2ba0-49ce-8842-f26edb941ad5" containerName="nova-manage" Dec 12 07:11:15 crc kubenswrapper[4867]: E1212 07:11:15.475303 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="dnsmasq-dns" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475362 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="dnsmasq-dns" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475658 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dfa393-ab5a-4aad-b7cc-cdc49d3d1da9" containerName="dnsmasq-dns" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475783 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-api" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.475996 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" containerName="nova-api-log" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.476211 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f25ce3-2ba0-49ce-8842-f26edb941ad5" containerName="nova-manage" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.477601 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.480202 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.480592 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.480882 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.487717 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.543568 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.544019 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.544327 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.544519 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.544770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5n8s\" (UniqueName: \"kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.545091 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.647683 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648001 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648135 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648590 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648779 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5n8s\" (UniqueName: \"kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648894 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.648986 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.652101 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.652206 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.660865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.662059 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.669675 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5n8s\" (UniqueName: \"kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s\") pod \"nova-api-0\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " pod="openstack/nova-api-0" Dec 12 07:11:15 crc kubenswrapper[4867]: I1212 07:11:15.811686 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:11:16 crc kubenswrapper[4867]: E1212 07:11:16.194748 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194 is running failed: container process not found" containerID="6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:11:16 crc kubenswrapper[4867]: E1212 07:11:16.195631 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194 is running failed: container process not found" containerID="6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:11:16 crc kubenswrapper[4867]: E1212 07:11:16.196261 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194 is running failed: container process not found" containerID="6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:11:16 crc kubenswrapper[4867]: E1212 07:11:16.196294 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerName="nova-scheduler-scheduler" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.264262 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.417016 4867 generic.go:334] "Generic (PLEG): container finished" podID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerID="6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" exitCode=0 Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.417084 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"09bd736a-175d-4a34-b0eb-714e10bd2207","Type":"ContainerDied","Data":"6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194"} Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.418108 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerStarted","Data":"4b978d6c6fa7d736b2906d3a81e5031001373012eaf81a8876153867f05c5c03"} Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.588824 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.667011 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh8rg\" (UniqueName: \"kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg\") pod \"09bd736a-175d-4a34-b0eb-714e10bd2207\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.667303 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle\") pod \"09bd736a-175d-4a34-b0eb-714e10bd2207\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.667437 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data\") pod \"09bd736a-175d-4a34-b0eb-714e10bd2207\" (UID: \"09bd736a-175d-4a34-b0eb-714e10bd2207\") " Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.671835 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg" (OuterVolumeSpecName: "kube-api-access-nh8rg") pod "09bd736a-175d-4a34-b0eb-714e10bd2207" (UID: "09bd736a-175d-4a34-b0eb-714e10bd2207"). InnerVolumeSpecName "kube-api-access-nh8rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.698684 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data" (OuterVolumeSpecName: "config-data") pod "09bd736a-175d-4a34-b0eb-714e10bd2207" (UID: "09bd736a-175d-4a34-b0eb-714e10bd2207"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.703930 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09bd736a-175d-4a34-b0eb-714e10bd2207" (UID: "09bd736a-175d-4a34-b0eb-714e10bd2207"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.770365 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.770414 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09bd736a-175d-4a34-b0eb-714e10bd2207-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.770432 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh8rg\" (UniqueName: \"kubernetes.io/projected/09bd736a-175d-4a34-b0eb-714e10bd2207-kube-api-access-nh8rg\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.849194 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a136e3a-f31d-44fb-9922-e73e7886592f" path="/var/lib/kubelet/pods/0a136e3a-f31d-44fb-9922-e73e7886592f/volumes" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.981062 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:33918->10.217.0.189:8775: read: connection reset by peer" Dec 12 07:11:16 crc kubenswrapper[4867]: I1212 07:11:16.981079 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": read tcp 10.217.0.2:33916->10.217.0.189:8775: read: connection reset by peer" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.430353 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"09bd736a-175d-4a34-b0eb-714e10bd2207","Type":"ContainerDied","Data":"09b19851156d33b364362c5216d737cf1bf604f291c379f0b7b8af6db1a9b9a1"} Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.430389 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.430680 4867 scope.go:117] "RemoveContainer" containerID="6da236a3d07c889a5cc40a39717f5eb10befddfce3f5013aa08ff4aede4d0194" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.435639 4867 generic.go:334] "Generic (PLEG): container finished" podID="4210192a-fbe4-4a91-981e-112ecf49a237" containerID="1671dbf245b4faa8dcbc3ca3be38cf020c46644501560453930a7a1ee4255c29" exitCode=0 Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.435746 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerDied","Data":"1671dbf245b4faa8dcbc3ca3be38cf020c46644501560453930a7a1ee4255c29"} Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.435778 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4210192a-fbe4-4a91-981e-112ecf49a237","Type":"ContainerDied","Data":"1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8"} Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.435791 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c36680369baa3d43c427dd00cfd10823febeda4f5a6767a3c3c8a9049327ed8" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.438569 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerStarted","Data":"b6413fee15f4b9393e64a3a1314ffb75d76cc3621c445d54a57ddbb7d9e78327"} Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.438600 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerStarted","Data":"f7cd3a86953862b945327d07440fe89481f1546ab42431b9a9517408da200482"} Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.476359 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.479733 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.479716008 podStartE2EDuration="2.479716008s" podCreationTimestamp="2025-12-12 07:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:11:17.465771784 +0000 UTC m=+1365.037153073" watchObservedRunningTime="2025-12-12 07:11:17.479716008 +0000 UTC m=+1365.051097297" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.493686 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.510370 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.527391 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:17 crc kubenswrapper[4867]: E1212 07:11:17.527860 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.527879 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" Dec 12 07:11:17 crc kubenswrapper[4867]: E1212 07:11:17.527894 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.527901 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" Dec 12 07:11:17 crc kubenswrapper[4867]: E1212 07:11:17.527934 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerName="nova-scheduler-scheduler" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.527940 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerName="nova-scheduler-scheduler" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.528117 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-log" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.528142 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" containerName="nova-metadata-metadata" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.528170 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" containerName="nova-scheduler-scheduler" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.528986 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.533091 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.535908 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586482 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data\") pod \"4210192a-fbe4-4a91-981e-112ecf49a237\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586522 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs\") pod \"4210192a-fbe4-4a91-981e-112ecf49a237\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586601 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs\") pod \"4210192a-fbe4-4a91-981e-112ecf49a237\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586640 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle\") pod \"4210192a-fbe4-4a91-981e-112ecf49a237\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586794 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbpgx\" (UniqueName: \"kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx\") pod \"4210192a-fbe4-4a91-981e-112ecf49a237\" (UID: \"4210192a-fbe4-4a91-981e-112ecf49a237\") " Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.586965 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs" (OuterVolumeSpecName: "logs") pod "4210192a-fbe4-4a91-981e-112ecf49a237" (UID: "4210192a-fbe4-4a91-981e-112ecf49a237"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.587111 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.587180 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz79s\" (UniqueName: \"kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.587244 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.587394 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4210192a-fbe4-4a91-981e-112ecf49a237-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.592678 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx" (OuterVolumeSpecName: "kube-api-access-tbpgx") pod "4210192a-fbe4-4a91-981e-112ecf49a237" (UID: "4210192a-fbe4-4a91-981e-112ecf49a237"). InnerVolumeSpecName "kube-api-access-tbpgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.621610 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data" (OuterVolumeSpecName: "config-data") pod "4210192a-fbe4-4a91-981e-112ecf49a237" (UID: "4210192a-fbe4-4a91-981e-112ecf49a237"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.631466 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4210192a-fbe4-4a91-981e-112ecf49a237" (UID: "4210192a-fbe4-4a91-981e-112ecf49a237"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.658555 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4210192a-fbe4-4a91-981e-112ecf49a237" (UID: "4210192a-fbe4-4a91-981e-112ecf49a237"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.688764 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz79s\" (UniqueName: \"kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.688861 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.688970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.689018 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.689032 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbpgx\" (UniqueName: \"kubernetes.io/projected/4210192a-fbe4-4a91-981e-112ecf49a237-kube-api-access-tbpgx\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.689042 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.689051 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4210192a-fbe4-4a91-981e-112ecf49a237-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.693522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.695179 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.717204 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz79s\" (UniqueName: \"kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s\") pod \"nova-scheduler-0\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " pod="openstack/nova-scheduler-0" Dec 12 07:11:17 crc kubenswrapper[4867]: I1212 07:11:17.855588 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.314115 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.449669 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"746ca9c4-7621-489c-86f2-628a9630039c","Type":"ContainerStarted","Data":"abbaa3c5bc249a78813a008ddc5a0d13a0c2722e39516639596ad3fbaa74c3e6"} Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.449936 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.490765 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.508567 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.525974 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.528619 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.532348 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.532701 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.552497 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.617017 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.617113 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.617373 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.617427 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2vc9\" (UniqueName: \"kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.617467 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.718885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.718969 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.719044 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.719066 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2vc9\" (UniqueName: \"kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.719088 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.795706 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.797859 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.802882 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.808485 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.815845 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2vc9\" (UniqueName: \"kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9\") pod \"nova-metadata-0\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " pod="openstack/nova-metadata-0" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.852041 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09bd736a-175d-4a34-b0eb-714e10bd2207" path="/var/lib/kubelet/pods/09bd736a-175d-4a34-b0eb-714e10bd2207/volumes" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.852889 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4210192a-fbe4-4a91-981e-112ecf49a237" path="/var/lib/kubelet/pods/4210192a-fbe4-4a91-981e-112ecf49a237/volumes" Dec 12 07:11:18 crc kubenswrapper[4867]: I1212 07:11:18.855538 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:11:19 crc kubenswrapper[4867]: I1212 07:11:19.316479 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:11:19 crc kubenswrapper[4867]: W1212 07:11:19.317487 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbb7b7c4_4c08_4f29_8e06_18af9f7422ae.slice/crio-282dc848c0d4f9b95110e2c42055c000a3d4db533e84c29c0e761f5689491da3 WatchSource:0}: Error finding container 282dc848c0d4f9b95110e2c42055c000a3d4db533e84c29c0e761f5689491da3: Status 404 returned error can't find the container with id 282dc848c0d4f9b95110e2c42055c000a3d4db533e84c29c0e761f5689491da3 Dec 12 07:11:19 crc kubenswrapper[4867]: I1212 07:11:19.462590 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerStarted","Data":"282dc848c0d4f9b95110e2c42055c000a3d4db533e84c29c0e761f5689491da3"} Dec 12 07:11:19 crc kubenswrapper[4867]: I1212 07:11:19.465109 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"746ca9c4-7621-489c-86f2-628a9630039c","Type":"ContainerStarted","Data":"aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164"} Dec 12 07:11:19 crc kubenswrapper[4867]: I1212 07:11:19.482393 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.482370902 podStartE2EDuration="2.482370902s" podCreationTimestamp="2025-12-12 07:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:11:19.481437809 +0000 UTC m=+1367.052819098" watchObservedRunningTime="2025-12-12 07:11:19.482370902 +0000 UTC m=+1367.053752171" Dec 12 07:11:20 crc kubenswrapper[4867]: I1212 07:11:20.476203 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerStarted","Data":"ee6e5dae4c9a11595f4e81ee6c481fdc2279b8c9a6898d55722fd9690e003f40"} Dec 12 07:11:20 crc kubenswrapper[4867]: I1212 07:11:20.476591 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerStarted","Data":"7af7ea533c4fbeeee8dc600c2255f6879f5ae8111689a0c44bb13bbf800546fe"} Dec 12 07:11:20 crc kubenswrapper[4867]: I1212 07:11:20.520391 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.52036994 podStartE2EDuration="2.52036994s" podCreationTimestamp="2025-12-12 07:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:11:20.503630126 +0000 UTC m=+1368.075011395" watchObservedRunningTime="2025-12-12 07:11:20.52036994 +0000 UTC m=+1368.091751209" Dec 12 07:11:22 crc kubenswrapper[4867]: I1212 07:11:22.856596 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 07:11:23 crc kubenswrapper[4867]: I1212 07:11:23.856811 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:11:23 crc kubenswrapper[4867]: I1212 07:11:23.856970 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 07:11:25 crc kubenswrapper[4867]: I1212 07:11:25.813004 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:11:25 crc kubenswrapper[4867]: I1212 07:11:25.813400 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 07:11:26 crc kubenswrapper[4867]: I1212 07:11:26.825383 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 07:11:26 crc kubenswrapper[4867]: I1212 07:11:26.825422 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 07:11:27 crc kubenswrapper[4867]: I1212 07:11:27.856609 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 07:11:27 crc kubenswrapper[4867]: I1212 07:11:27.892448 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 07:11:28 crc kubenswrapper[4867]: I1212 07:11:28.576852 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 07:11:28 crc kubenswrapper[4867]: I1212 07:11:28.856646 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 07:11:28 crc kubenswrapper[4867]: I1212 07:11:28.856712 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 07:11:29 crc kubenswrapper[4867]: I1212 07:11:29.823518 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 12 07:11:29 crc kubenswrapper[4867]: I1212 07:11:29.871504 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 07:11:29 crc kubenswrapper[4867]: I1212 07:11:29.871517 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.629636 4867 generic.go:334] "Generic (PLEG): container finished" podID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerID="11d1470839b0af49726e251b1c18e90b839155db336892a680b70848ec7eceaf" exitCode=137 Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.629734 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerDied","Data":"11d1470839b0af49726e251b1c18e90b839155db336892a680b70848ec7eceaf"} Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.822810 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.823680 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.824859 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 07:11:35 crc kubenswrapper[4867]: I1212 07:11:35.830345 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.253907 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.399908 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400091 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dkh6\" (UniqueName: \"kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400127 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400200 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400827 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400905 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.400948 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.401109 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd\") pod \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\" (UID: \"a3950b22-ca71-44c2-89e6-1209a53d0cd3\") " Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.401210 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.401829 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.402018 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.402047 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3950b22-ca71-44c2-89e6-1209a53d0cd3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.414816 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6" (OuterVolumeSpecName: "kube-api-access-9dkh6") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "kube-api-access-9dkh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.414925 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts" (OuterVolumeSpecName: "scripts") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.439146 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.456112 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.487142 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.504578 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.504632 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dkh6\" (UniqueName: \"kubernetes.io/projected/a3950b22-ca71-44c2-89e6-1209a53d0cd3-kube-api-access-9dkh6\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.504645 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.504654 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.504664 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.508490 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data" (OuterVolumeSpecName: "config-data") pod "a3950b22-ca71-44c2-89e6-1209a53d0cd3" (UID: "a3950b22-ca71-44c2-89e6-1209a53d0cd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.607052 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3950b22-ca71-44c2-89e6-1209a53d0cd3-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.646803 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3950b22-ca71-44c2-89e6-1209a53d0cd3","Type":"ContainerDied","Data":"2ae63e7053d55c1cd068e85c905d24ac70b4bc8581d9650b5919da596ed78aa2"} Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.646870 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.646877 4867 scope.go:117] "RemoveContainer" containerID="11d1470839b0af49726e251b1c18e90b839155db336892a680b70848ec7eceaf" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.648194 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.657432 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.685393 4867 scope.go:117] "RemoveContainer" containerID="253e8f652d0053f45aa89e8099a1d53ff79e1d5dfbb7086e369f0fb29d2905ef" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.735655 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.737347 4867 scope.go:117] "RemoveContainer" containerID="1cbd7ae07f31048ed723481fba0f7b7f22a511f8cafc77bbfecf93da165d05a8" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.766578 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.774361 4867 scope.go:117] "RemoveContainer" containerID="715f3afdac23fd09e49196492e04351d6f822ad0a7b572801de345b914b2a1c1" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783177 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:36 crc kubenswrapper[4867]: E1212 07:11:36.783632 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-central-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783657 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-central-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: E1212 07:11:36.783683 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="proxy-httpd" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783691 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="proxy-httpd" Dec 12 07:11:36 crc kubenswrapper[4867]: E1212 07:11:36.783715 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-notification-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783723 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-notification-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: E1212 07:11:36.783740 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="sg-core" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783746 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="sg-core" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783912 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-notification-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783946 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="sg-core" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783963 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="ceilometer-central-agent" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.783976 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" containerName="proxy-httpd" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.786067 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.801849 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.819934 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.820782 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.852863 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.873339 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3950b22-ca71-44c2-89e6-1209a53d0cd3" path="/var/lib/kubelet/pods/a3950b22-ca71-44c2-89e6-1209a53d0cd3/volumes" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.925765 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.925830 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.925869 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.925892 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.925921 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.926007 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbm8x\" (UniqueName: \"kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.926022 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:36 crc kubenswrapper[4867]: I1212 07:11:36.926063 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.027831 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.027899 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.027942 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028056 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbm8x\" (UniqueName: \"kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028081 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028137 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028199 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028284 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.028992 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.029057 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.032101 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.032866 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.033523 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.041956 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.050311 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.057993 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbm8x\" (UniqueName: \"kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x\") pod \"ceilometer-0\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.158402 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:11:37 crc kubenswrapper[4867]: I1212 07:11:37.638889 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:11:37 crc kubenswrapper[4867]: W1212 07:11:37.659352 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a9d745d_1642_48f4_b1ee_e491f578757e.slice/crio-70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082 WatchSource:0}: Error finding container 70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082: Status 404 returned error can't find the container with id 70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082 Dec 12 07:11:38 crc kubenswrapper[4867]: I1212 07:11:38.665879 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerStarted","Data":"70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082"} Dec 12 07:11:38 crc kubenswrapper[4867]: I1212 07:11:38.860904 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 07:11:38 crc kubenswrapper[4867]: I1212 07:11:38.862099 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 07:11:38 crc kubenswrapper[4867]: I1212 07:11:38.866476 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 07:11:39 crc kubenswrapper[4867]: I1212 07:11:39.680541 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerStarted","Data":"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a"} Dec 12 07:11:39 crc kubenswrapper[4867]: I1212 07:11:39.687677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 07:11:42 crc kubenswrapper[4867]: I1212 07:11:42.706477 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerStarted","Data":"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474"} Dec 12 07:11:42 crc kubenswrapper[4867]: I1212 07:11:42.707303 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerStarted","Data":"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece"} Dec 12 07:11:44 crc kubenswrapper[4867]: I1212 07:11:44.729363 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerStarted","Data":"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b"} Dec 12 07:11:44 crc kubenswrapper[4867]: I1212 07:11:44.730055 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 07:11:44 crc kubenswrapper[4867]: I1212 07:11:44.760396 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.631357762 podStartE2EDuration="8.760372965s" podCreationTimestamp="2025-12-12 07:11:36 +0000 UTC" firstStartedPulling="2025-12-12 07:11:37.662258522 +0000 UTC m=+1385.233639791" lastFinishedPulling="2025-12-12 07:11:43.791273725 +0000 UTC m=+1391.362654994" observedRunningTime="2025-12-12 07:11:44.753257295 +0000 UTC m=+1392.324638584" watchObservedRunningTime="2025-12-12 07:11:44.760372965 +0000 UTC m=+1392.331754234" Dec 12 07:11:58 crc kubenswrapper[4867]: I1212 07:11:58.989217 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:11:58 crc kubenswrapper[4867]: I1212 07:11:58.989801 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:12:07 crc kubenswrapper[4867]: I1212 07:12:07.166544 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.465317 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.475886 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="479f0ca7-1408-41d4-abaf-6034d0802100" containerName="openstackclient" containerID="cri-o://48b3ddfeb9f53136e00be49e14a43f05d1fd34eb01693af6aa1fd779aa73a984" gracePeriod=2 Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.503949 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.604250 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.604527 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" containerID="cri-o://7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" gracePeriod=30 Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.604919 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="openstack-network-exporter" containerID="cri-o://0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad" gracePeriod=30 Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.695212 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.824977 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:12:26 crc kubenswrapper[4867]: I1212 07:12:26.826162 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="openstack-network-exporter" containerID="cri-o://48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098" gracePeriod=300 Dec 12 07:12:27 crc kubenswrapper[4867]: E1212 07:12:26.919742 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:27 crc kubenswrapper[4867]: E1212 07:12:26.919808 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data podName:65ad43b4-b2fa-4089-b471-66fbae81fbf2 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:27.419789484 +0000 UTC m=+1434.991170753 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data") pod "rabbitmq-cell1-server-0" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2") : configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.225521 4867 generic.go:334] "Generic (PLEG): container finished" podID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerID="48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098" exitCode=2 Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.227600 4867 generic.go:334] "Generic (PLEG): container finished" podID="00269882-fe60-4b26-8d5c-1b34b014b191" containerID="0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad" exitCode=2 Dec 12 07:12:27 crc kubenswrapper[4867]: E1212 07:12:27.462568 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:27 crc kubenswrapper[4867]: E1212 07:12:27.462634 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data podName:65ad43b4-b2fa-4089-b471-66fbae81fbf2 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:28.462620282 +0000 UTC m=+1436.034001551 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data") pod "rabbitmq-cell1-server-0" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2") : configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.839483 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:27 crc kubenswrapper[4867]: E1212 07:12:27.839763 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479f0ca7-1408-41d4-abaf-6034d0802100" containerName="openstackclient" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.839776 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="479f0ca7-1408-41d4-abaf-6034d0802100" containerName="openstackclient" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.840054 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="479f0ca7-1408-41d4-abaf-6034d0802100" containerName="openstackclient" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.853644 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerDied","Data":"48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098"} Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.853698 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.853715 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.854618 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.854633 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855347 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855379 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855706 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerDied","Data":"0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad"} Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855735 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855746 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855758 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6lmsp"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855770 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6lmsp"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.855781 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.857805 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.861002 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.865239 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9dbsw"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.865257 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9dbsw"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.865892 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.861079 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.859968 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="dnsmasq-dns" containerID="cri-o://5b998256ed7e715606082bd623b651120a069bd1bdab372dad3314ef077aeaaf" gracePeriod=10 Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867283 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-7lxjs"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867305 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-7lxjs"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867318 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867327 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7fnd8"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867339 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7fnd8"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867350 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.867421 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.868142 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-ptmqx"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.868163 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-ptmqx"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.869215 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.877172 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.885962 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886028 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886087 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whk44\" (UniqueName: \"kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886111 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886130 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fg6n\" (UniqueName: \"kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886150 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrlk6\" (UniqueName: \"kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886201 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886282 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqfsk\" (UniqueName: \"kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886306 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886350 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2rxk\" (UniqueName: \"kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.886377 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb2k7\" (UniqueName: \"kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.917607 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-bxvfm"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.930658 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-bxvfm"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.965335 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.976085 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.984720 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987540 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb2k7\" (UniqueName: \"kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987583 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987614 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987636 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987683 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfbs\" (UniqueName: \"kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987716 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whk44\" (UniqueName: \"kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987782 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987804 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fg6n\" (UniqueName: \"kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987823 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrlk6\" (UniqueName: \"kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987840 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987872 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987959 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqfsk\" (UniqueName: \"kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.987981 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.988020 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2rxk\" (UniqueName: \"kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.988630 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.989078 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.989521 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.989905 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.992336 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:27 crc kubenswrapper[4867]: I1212 07:12:27.992546 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.025630 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.029317 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-server" containerID="cri-o://ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032099 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-updater" containerID="cri-o://572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032267 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="rsync" containerID="cri-o://3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032309 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-expirer" containerID="cri-o://517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032343 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-updater" containerID="cri-o://b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032376 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-auditor" containerID="cri-o://b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032404 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-replicator" containerID="cri-o://6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032434 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-server" containerID="cri-o://1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032486 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-reaper" containerID="cri-o://6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032521 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-auditor" containerID="cri-o://bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032550 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-replicator" containerID="cri-o://6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032579 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-server" containerID="cri-o://a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.034190 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-auditor" containerID="cri-o://2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.032624 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-replicator" containerID="cri-o://f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.051985 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="swift-recon-cron" containerID="cri-o://09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.055523 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whk44\" (UniqueName: \"kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44\") pod \"cinderb8be-account-delete-kvnwc\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.058999 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2rxk\" (UniqueName: \"kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk\") pod \"neutron29f2-account-delete-77p2l\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.061707 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrlk6\" (UniqueName: \"kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6\") pod \"barbican74a1-account-delete-m9cms\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.063214 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb2k7\" (UniqueName: \"kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7\") pod \"placementc9b0-account-delete-5g27c\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.069450 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="ovsdbserver-nb" containerID="cri-o://d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" gracePeriod=299 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.073941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fg6n\" (UniqueName: \"kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n\") pod \"novaapida12-account-delete-nl9mn\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.091985 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqfsk\" (UniqueName: \"kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk\") pod \"glancea6d9-account-delete-zqpf9\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.101286 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.101445 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfbs\" (UniqueName: \"kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.102764 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.102808 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.135085 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.142471 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="openstack-network-exporter" containerID="cri-o://ef733b2c26cb5cf662dfa06f03ae9d9a2ad1b9bfc775896fa039f832b4e8f2b0" gracePeriod=300 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.180853 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfbs\" (UniqueName: \"kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs\") pod \"novacell0896e-account-delete-9tg4b\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.208770 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.287881 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.288475 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-68ccfcbfb4-q47vm" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-log" containerID="cri-o://7d0afa62d13fd39fc3b3ba841aff538f37e9ec010eeac42f3c84ad166e59f590" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.288979 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-68ccfcbfb4-q47vm" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-api" containerID="cri-o://a6b387c7b3b91393496b3df6c1f76038666dc95195810690a8e0c383d5e2aa2d" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.324633 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="ovsdbserver-sb" containerID="cri-o://a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" gracePeriod=300 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.340033 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.340326 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-92rzj" podUID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" containerName="openstack-network-exporter" containerID="cri-o://c4b46d15a056004b20724580a9403bb36eec9ea66794fe8c2015a7db598e5ae8" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.436060 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.509373 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.509739 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6796c56757-d2snl" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-api" containerID="cri-o://83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.509892 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6796c56757-d2snl" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-httpd" containerID="cri-o://7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.544938 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.545509 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-log" containerID="cri-o://506b0e61cb33a6fbdd013a5cdbf2ecb8a9e984240941a905449f20e6d9bda9a1" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.545699 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-httpd" containerID="cri-o://f164d088b657a841b64a09bed96dee589e4f7130244d17533d2c9a7d6fdb6dab" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: E1212 07:12:28.553048 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:28 crc kubenswrapper[4867]: E1212 07:12:28.553111 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data podName:65ad43b4-b2fa-4089-b471-66fbae81fbf2 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:30.553096148 +0000 UTC m=+1438.124477417 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data") pod "rabbitmq-cell1-server-0" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2") : configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:28 crc kubenswrapper[4867]: E1212 07:12:28.555069 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 12 07:12:28 crc kubenswrapper[4867]: E1212 07:12:28.555110 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data podName:00532992-54f5-4a24-b0e4-eb5512e08753 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:29.055099399 +0000 UTC m=+1436.626480668 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data") pod "rabbitmq-server-0" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753") : configmap "rabbitmq-config-data" not found Dec 12 07:12:28 crc kubenswrapper[4867]: E1212 07:12:28.555719 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e39305a_805b_4e9f_ac88_e85e8007409f.slice/crio-48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-conmon-b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00269882_fe60_4b26_8d5c_1b34b014b191.slice/crio-conmon-0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-conmon-6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-conmon-2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00269882_fe60_4b26_8d5c_1b34b014b191.slice/crio-0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e39305a_805b_4e9f_ac88_e85e8007409f.slice/crio-conmon-48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-conmon-6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod786b1010_d0a8_4c94_a42c_99226ab568be.slice/crio-conmon-572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04569a7e_a52f_4c83_8d2d_e2d6c1ac6d6f.slice/crio-a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df.scope\": RecentStats: unable to find data in memory cache]" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.583703 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9c5ps"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.615403 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9c5ps"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.654335 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zjmsk"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.689062 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zjmsk"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.698696 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.698925 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="cinder-scheduler" containerID="cri-o://8cd180ed7cb6b465bdfd0105a35071f5be7a9810e30b308406669748fe930aab" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.699369 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="probe" containerID="cri-o://82e68b2b9f69ed30e1eee05edc30a399c219b479de169d6d7373d0b615332b11" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.713413 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.713716 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-log" containerID="cri-o://e0e23c15cb95fadeb8c5fb6992fb31533b6edb8b32c8f954630c0f55ff2e3c91" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.714099 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-httpd" containerID="cri-o://4a8ad42ecdb710efacb437807e79af57dea08fcccf32bfdc46ca2fca80f02ba9" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.725084 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.725384 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api-log" containerID="cri-o://f3626d4c1ef6132f0076b910eab89601c61a779670954414124b9e2ed04cba99" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.725552 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" containerID="cri-o://384bb1fccf6418383d28e93ad17cb3f3d26be31108864c9e588482c7982d8d1c" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.772049 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.772351 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-67c5f5d445-rd5dx" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-httpd" containerID="cri-o://bc30a378f2f47b54d0e25ab70eab05c43f7eb5652e82cd1c2e5b19b297703f21" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.772776 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-67c5f5d445-rd5dx" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-server" containerID="cri-o://dc60c646092d5fd6e48768335070dbf4389a632716a9616ea70be7334134e290" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.781439 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.781675 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker-log" containerID="cri-o://1499f0119bb5be4db9007f30932c9a39e2288f464937e17ccb75934aea9e9680" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.781998 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker" containerID="cri-o://0482c34d8c8358e81c2f8bc9ca4911f7c07f4fe5e94574f9a0a5812a8529427f" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.805875 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.806763 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener-log" containerID="cri-o://14a5c4be4116f44371426de020f2bddb11420dd71b7ecd0f938dff04595d0193" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.807133 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener" containerID="cri-o://23284a0c2e794f7efbc84dbb67df86a5355b8c4e5689bfc88bd6bad7d0096872" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.820462 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.820803 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" containerID="cri-o://e77048aeb4e616861c5d278c417e28346695e6edbd2cfceba516f723cc72bdf9" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.820953 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" containerID="cri-o://17557e0767d1d09b0c76be9d8ba0d9223b9b9565bc8ce63ba2fb3d24cdeacbe1" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.832251 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.832537 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" containerID="cri-o://7af7ea533c4fbeeee8dc600c2255f6879f5ae8111689a0c44bb13bbf800546fe" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.833414 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" containerID="cri-o://ee6e5dae4c9a11595f4e81ee6c481fdc2279b8c9a6898d55722fd9690e003f40" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.855180 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1432b4ca-14f2-4044-a58c-d93ff9930e17" path="/var/lib/kubelet/pods/1432b4ca-14f2-4044-a58c-d93ff9930e17/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.856587 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f78a88a-c256-4e35-8f64-1c155650cca6" path="/var/lib/kubelet/pods/3f78a88a-c256-4e35-8f64-1c155650cca6/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.857794 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469654c1-0303-4127-a61c-411d65a09cd2" path="/var/lib/kubelet/pods/469654c1-0303-4127-a61c-411d65a09cd2/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.859312 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63341abf-5801-4239-8370-4c0732c95e3b" path="/var/lib/kubelet/pods/63341abf-5801-4239-8370-4c0732c95e3b/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.860317 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68801d28-d85f-46b6-86cd-debaf6dab7ca" path="/var/lib/kubelet/pods/68801d28-d85f-46b6-86cd-debaf6dab7ca/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.861017 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7299cd15-ce08-48f3-be09-c3a70901c4aa" path="/var/lib/kubelet/pods/7299cd15-ce08-48f3-be09-c3a70901c4aa/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.861786 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f25ce3-2ba0-49ce-8842-f26edb941ad5" path="/var/lib/kubelet/pods/91f25ce3-2ba0-49ce-8842-f26edb941ad5/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.877539 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3a5237b-ffff-43b1-a921-68affabec6cd" path="/var/lib/kubelet/pods/a3a5237b-ffff-43b1-a921-68affabec6cd/volumes" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878511 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878560 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878579 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b624l"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878595 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b624l"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878615 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2010-account-create-update-lfxbp"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.878896 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-log" containerID="cri-o://f7cd3a86953862b945327d07440fe89481f1546ab42431b9a9517408da200482" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.879181 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-api" containerID="cri-o://b6413fee15f4b9393e64a3a1314ffb75d76cc3621c445d54a57ddbb7d9e78327" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.885281 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2010-account-create-update-lfxbp"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.899484 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.904664 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.904929 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://327346d748281830108225a6a9e113b8722507f57f85ac4bbb3af491d43078a5" gracePeriod=30 Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.910476 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.989250 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.989523 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.999312 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:12:28 crc kubenswrapper[4867]: I1212 07:12:28.999563 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="746ca9c4-7621-489c-86f2-628a9630039c" containerName="nova-scheduler-scheduler" containerID="cri-o://aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" gracePeriod=30 Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.089187 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.089289 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data podName:00532992-54f5-4a24-b0e4-eb5512e08753 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:30.089272247 +0000 UTC m=+1437.660653516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data") pod "rabbitmq-server-0" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753") : configmap "rabbitmq-config-data" not found Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.170809 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="rabbitmq" containerID="cri-o://234415f3af38fbe83a304c1655d565aaee048e266eab8c74960cba70dcab3267" gracePeriod=604800 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.170989 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.204111 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.245747 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.273132 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" containerID="cri-o://fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" gracePeriod=29 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.298754 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.337416 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f is running failed: container process not found" containerID="d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.337801 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f is running failed: container process not found" containerID="d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.339526 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f is running failed: container process not found" containerID="d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.339600 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="ovsdbserver-nb" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.350885 4867 generic.go:334] "Generic (PLEG): container finished" podID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerID="5b998256ed7e715606082bd623b651120a069bd1bdab372dad3314ef077aeaaf" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.351043 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" event={"ID":"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8","Type":"ContainerDied","Data":"5b998256ed7e715606082bd623b651120a069bd1bdab372dad3314ef077aeaaf"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.351081 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" event={"ID":"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8","Type":"ContainerDied","Data":"5bd17181a1f2fd4f740b3706415ba0cb8648d14a66fe5f0309d85a79d3e28d31"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.351135 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bd17181a1f2fd4f740b3706415ba0cb8648d14a66fe5f0309d85a79d3e28d31" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.353142 4867 generic.go:334] "Generic (PLEG): container finished" podID="479f0ca7-1408-41d4-abaf-6034d0802100" containerID="48b3ddfeb9f53136e00be49e14a43f05d1fd34eb01693af6aa1fd779aa73a984" exitCode=137 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.373168 4867 generic.go:334] "Generic (PLEG): container finished" podID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerID="506b0e61cb33a6fbdd013a5cdbf2ecb8a9e984240941a905449f20e6d9bda9a1" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.373258 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerDied","Data":"506b0e61cb33a6fbdd013a5cdbf2ecb8a9e984240941a905449f20e6d9bda9a1"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.376472 4867 generic.go:334] "Generic (PLEG): container finished" podID="765da323-2d60-4494-97bc-f699d9bf8661" containerID="7d0afa62d13fd39fc3b3ba841aff538f37e9ec010eeac42f3c84ad166e59f590" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.376536 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerDied","Data":"7d0afa62d13fd39fc3b3ba841aff538f37e9ec010eeac42f3c84ad166e59f590"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.381801 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-92rzj_c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba/openstack-network-exporter/0.log" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.381844 4867 generic.go:334] "Generic (PLEG): container finished" podID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" containerID="c4b46d15a056004b20724580a9403bb36eec9ea66794fe8c2015a7db598e5ae8" exitCode=2 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.381906 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-92rzj" event={"ID":"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba","Type":"ContainerDied","Data":"c4b46d15a056004b20724580a9403bb36eec9ea66794fe8c2015a7db598e5ae8"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.394385 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f/ovsdbserver-sb/0.log" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.394434 4867 generic.go:334] "Generic (PLEG): container finished" podID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerID="ef733b2c26cb5cf662dfa06f03ae9d9a2ad1b9bfc775896fa039f832b4e8f2b0" exitCode=2 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.394475 4867 generic.go:334] "Generic (PLEG): container finished" podID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerID="a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.394595 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerDied","Data":"ef733b2c26cb5cf662dfa06f03ae9d9a2ad1b9bfc775896fa039f832b4e8f2b0"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.394633 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerDied","Data":"a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.399645 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.406309 4867 generic.go:334] "Generic (PLEG): container finished" podID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerID="e0e23c15cb95fadeb8c5fb6992fb31533b6edb8b32c8f954630c0f55ff2e3c91" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.406387 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerDied","Data":"e0e23c15cb95fadeb8c5fb6992fb31533b6edb8b32c8f954630c0f55ff2e3c91"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.408819 4867 generic.go:334] "Generic (PLEG): container finished" podID="163cec12-d32c-4be9-aee0-55703387f2db" containerID="f3626d4c1ef6132f0076b910eab89601c61a779670954414124b9e2ed04cba99" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.408875 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerDied","Data":"f3626d4c1ef6132f0076b910eab89601c61a779670954414124b9e2ed04cba99"} Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.409398 4867 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 12 07:12:29 crc kubenswrapper[4867]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 12 07:12:29 crc kubenswrapper[4867]: + source /usr/local/bin/container-scripts/functions Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNBridge=br-int Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNRemote=tcp:localhost:6642 Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNEncapType=geneve Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNAvailabilityZones= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ EnableChassisAsGateway=true Dec 12 07:12:29 crc kubenswrapper[4867]: ++ PhysicalNetworks= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNHostName= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 12 07:12:29 crc kubenswrapper[4867]: ++ ovs_dir=/var/lib/openvswitch Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 12 07:12:29 crc kubenswrapper[4867]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + cleanup_ovsdb_server_semaphore Dec 12 07:12:29 crc kubenswrapper[4867]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 12 07:12:29 crc kubenswrapper[4867]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-9k5zx" message=< Dec 12 07:12:29 crc kubenswrapper[4867]: Exiting ovsdb-server (5) [ OK ] Dec 12 07:12:29 crc kubenswrapper[4867]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 12 07:12:29 crc kubenswrapper[4867]: + source /usr/local/bin/container-scripts/functions Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNBridge=br-int Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNRemote=tcp:localhost:6642 Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNEncapType=geneve Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNAvailabilityZones= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ EnableChassisAsGateway=true Dec 12 07:12:29 crc kubenswrapper[4867]: ++ PhysicalNetworks= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNHostName= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 12 07:12:29 crc kubenswrapper[4867]: ++ ovs_dir=/var/lib/openvswitch Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 12 07:12:29 crc kubenswrapper[4867]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + cleanup_ovsdb_server_semaphore Dec 12 07:12:29 crc kubenswrapper[4867]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 12 07:12:29 crc kubenswrapper[4867]: > Dec 12 07:12:29 crc kubenswrapper[4867]: E1212 07:12:29.409440 4867 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 12 07:12:29 crc kubenswrapper[4867]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 12 07:12:29 crc kubenswrapper[4867]: + source /usr/local/bin/container-scripts/functions Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNBridge=br-int Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNRemote=tcp:localhost:6642 Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNEncapType=geneve Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNAvailabilityZones= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ EnableChassisAsGateway=true Dec 12 07:12:29 crc kubenswrapper[4867]: ++ PhysicalNetworks= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ OVNHostName= Dec 12 07:12:29 crc kubenswrapper[4867]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 12 07:12:29 crc kubenswrapper[4867]: ++ ovs_dir=/var/lib/openvswitch Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 12 07:12:29 crc kubenswrapper[4867]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 12 07:12:29 crc kubenswrapper[4867]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + sleep 0.5 Dec 12 07:12:29 crc kubenswrapper[4867]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 12 07:12:29 crc kubenswrapper[4867]: + cleanup_ovsdb_server_semaphore Dec 12 07:12:29 crc kubenswrapper[4867]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 12 07:12:29 crc kubenswrapper[4867]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 12 07:12:29 crc kubenswrapper[4867]: > pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" containerID="cri-o://969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.409495 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" containerID="cri-o://969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" gracePeriod=29 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.414847 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5e39305a-805b-4e9f-ac88-e85e8007409f/ovsdbserver-nb/0.log" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.414898 4867 generic.go:334] "Generic (PLEG): container finished" podID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerID="d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" exitCode=143 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.415104 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerDied","Data":"d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.415180 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5e39305a-805b-4e9f-ac88-e85e8007409f","Type":"ContainerDied","Data":"055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.415196 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="055f6dccd0b57e982dff3c7aed14dbc2668c05e115326aae9135d5bfc96cdbb6" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.425272 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428421 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428645 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428768 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428862 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428959 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429064 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429164 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429361 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429462 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429771 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.429854 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.430012 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.430088 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab" exitCode=0 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.428751 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.430845 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.430935 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431019 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431150 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431258 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431352 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431546 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431637 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431718 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.431791 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab"} Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.440019 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.452422 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.459428 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5e39305a-805b-4e9f-ac88-e85e8007409f/ovsdbserver-nb/0.log" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.459806 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.499880 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.499912 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.499942 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.499969 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500057 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500160 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500200 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500239 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br7h8\" (UniqueName: \"kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500266 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500290 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8hh8\" (UniqueName: \"kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500313 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500328 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc\") pod \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\" (UID: \"2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500342 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.500358 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle\") pod \"5e39305a-805b-4e9f-ac88-e85e8007409f\" (UID: \"5e39305a-805b-4e9f-ac88-e85e8007409f\") " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.502406 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts" (OuterVolumeSpecName: "scripts") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.504841 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config" (OuterVolumeSpecName: "config") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.512056 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.521873 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="rabbitmq" containerID="cri-o://02be4e6b1d82467e195a35fda03f0c2a9e4fbba01b3f18526ddc7beb6bb54785" gracePeriod=604800 Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.530761 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8" (OuterVolumeSpecName: "kube-api-access-c8hh8") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "kube-api-access-c8hh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.542192 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8" (OuterVolumeSpecName: "kube-api-access-br7h8") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "kube-api-access-br7h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.542357 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.588903 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602717 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602750 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602765 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br7h8\" (UniqueName: \"kubernetes.io/projected/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-kube-api-access-br7h8\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602795 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602808 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8hh8\" (UniqueName: \"kubernetes.io/projected/5e39305a-805b-4e9f-ac88-e85e8007409f-kube-api-access-c8hh8\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602820 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e39305a-805b-4e9f-ac88-e85e8007409f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.602833 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.611850 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config" (OuterVolumeSpecName: "config") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.643288 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.647738 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.657145 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.667974 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.674653 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.691252 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" (UID: "2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.704954 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.704990 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.705008 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.705028 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.705039 4867 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.705051 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.705064 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.739533 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "5e39305a-805b-4e9f-ac88-e85e8007409f" (UID: "5e39305a-805b-4e9f-ac88-e85e8007409f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.806131 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.808418 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e39305a-805b-4e9f-ac88-e85e8007409f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.894297 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:29 crc kubenswrapper[4867]: I1212 07:12:29.903079 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:29 crc kubenswrapper[4867]: W1212 07:12:29.908711 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c8e566d_e439_4e4b_bb80_759e9c4d7f89.slice/crio-95bcf6dd988818d74803c0b4f049b2ce430384f2bf06f2d6c67e3c6a960c0d61 WatchSource:0}: Error finding container 95bcf6dd988818d74803c0b4f049b2ce430384f2bf06f2d6c67e3c6a960c0d61: Status 404 returned error can't find the container with id 95bcf6dd988818d74803c0b4f049b2ce430384f2bf06f2d6c67e3c6a960c0d61 Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.046812 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.114909 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.114975 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data podName:00532992-54f5-4a24-b0e4-eb5512e08753 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:32.114958872 +0000 UTC m=+1439.686340141 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data") pod "rabbitmq-server-0" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753") : configmap "rabbitmq-config-data" not found Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.196530 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:30 crc kubenswrapper[4867]: W1212 07:12:30.198891 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd36e355_181a_4042_a049_aeda5ec0770f.slice/crio-8413c1fe30ad84dcb710d7a144e960c9c39920973b26eeabf39fdc387b07154c WatchSource:0}: Error finding container 8413c1fe30ad84dcb710d7a144e960c9c39920973b26eeabf39fdc387b07154c: Status 404 returned error can't find the container with id 8413c1fe30ad84dcb710d7a144e960c9c39920973b26eeabf39fdc387b07154c Dec 12 07:12:30 crc kubenswrapper[4867]: W1212 07:12:30.205889 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice/crio-9dc4fd28ec0f317f12400cdbe7542efe8ef98a5fdf7d59d1a6cd1b0e4a02c823 WatchSource:0}: Error finding container 9dc4fd28ec0f317f12400cdbe7542efe8ef98a5fdf7d59d1a6cd1b0e4a02c823: Status 404 returned error can't find the container with id 9dc4fd28ec0f317f12400cdbe7542efe8ef98a5fdf7d59d1a6cd1b0e4a02c823 Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.215931 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.298415 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:30 crc kubenswrapper[4867]: W1212 07:12:30.311869 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7dc7ba4_cc77_4569_ab0f_4bff75ab6472.slice/crio-428e5cb15516a50e79f56f0b9879e83f4c6464aa13842b57b5ba94dfa10f28d1 WatchSource:0}: Error finding container 428e5cb15516a50e79f56f0b9879e83f4c6464aa13842b57b5ba94dfa10f28d1: Status 404 returned error can't find the container with id 428e5cb15516a50e79f56f0b9879e83f4c6464aa13842b57b5ba94dfa10f28d1 Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.374996 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.383310 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.388516 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.388574 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.442209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerStarted","Data":"020f4a25f965ce85a1f6f8b78cc434900900698068d74a3b8073045729e7fd68"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.443471 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapida12-account-delete-nl9mn" event={"ID":"3c8e566d-e439-4e4b-bb80-759e9c4d7f89","Type":"ContainerStarted","Data":"95bcf6dd988818d74803c0b4f049b2ce430384f2bf06f2d6c67e3c6a960c0d61"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.445010 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerStarted","Data":"8413c1fe30ad84dcb710d7a144e960c9c39920973b26eeabf39fdc387b07154c"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.446620 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerStarted","Data":"c79730947610ada694194883b4630e38707000fca4d36cb10a6f0ad33476e446"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.448126 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0896e-account-delete-9tg4b" event={"ID":"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472","Type":"ContainerStarted","Data":"428e5cb15516a50e79f56f0b9879e83f4c6464aa13842b57b5ba94dfa10f28d1"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.449310 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerStarted","Data":"9dc4fd28ec0f317f12400cdbe7542efe8ef98a5fdf7d59d1a6cd1b0e4a02c823"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.450898 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerStarted","Data":"22faa0c03d72537d3b22c09966e75ec38e80847b6e5abd135c1b38a198d21c52"} Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.450949 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcd6f8f8f-whlhl" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.451077 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.501052 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.510465 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcd6f8f8f-whlhl"] Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.517766 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.524731 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.530950 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="galera" containerID="cri-o://49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" gracePeriod=29 Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.625133 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.625982 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data podName:65ad43b4-b2fa-4089-b471-66fbae81fbf2 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.625960253 +0000 UTC m=+1442.197341522 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data") pod "rabbitmq-cell1-server-0" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2") : configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.870085 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" path="/var/lib/kubelet/pods/2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8/volumes" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.870835 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" path="/var/lib/kubelet/pods/5e39305a-805b-4e9f-ac88-e85e8007409f/volumes" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.871848 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5b5420-43c8-4f14-b059-42eb6c7b312f" path="/var/lib/kubelet/pods/8a5b5420-43c8-4f14-b059-42eb6c7b312f/volumes" Dec 12 07:12:30 crc kubenswrapper[4867]: I1212 07:12:30.873183 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc646b27-e211-4705-91d4-e97fe112855a" path="/var/lib/kubelet/pods/dc646b27-e211-4705-91d4-e97fe112855a/volumes" Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.891493 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df is running failed: container process not found" containerID="a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.892351 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df is running failed: container process not found" containerID="a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.896537 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df is running failed: container process not found" containerID="a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 12 07:12:30 crc kubenswrapper[4867]: E1212 07:12:30.896584 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="ovsdbserver-sb" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.216075 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-92rzj_c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba/openstack-network-exporter/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.216499 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.316239 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f/ovsdbserver-sb/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.316312 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.319364 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355285 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355370 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd9wz\" (UniqueName: \"kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355464 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355488 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355525 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.355549 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config\") pod \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\" (UID: \"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.356950 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.357361 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.357903 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config" (OuterVolumeSpecName: "config") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.361844 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz" (OuterVolumeSpecName: "kube-api-access-pd9wz") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "kube-api-access-pd9wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.393526 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.436615 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" (UID: "c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.458837 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.458899 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.458966 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.458991 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle\") pod \"479f0ca7-1408-41d4-abaf-6034d0802100\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459032 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459063 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459095 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5pbb\" (UniqueName: \"kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb\") pod \"479f0ca7-1408-41d4-abaf-6034d0802100\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459117 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459156 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret\") pod \"479f0ca7-1408-41d4-abaf-6034d0802100\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459181 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459219 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config\") pod \"479f0ca7-1408-41d4-abaf-6034d0802100\" (UID: \"479f0ca7-1408-41d4-abaf-6034d0802100\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459369 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n5tf\" (UniqueName: \"kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf\") pod \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\" (UID: \"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f\") " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459772 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459788 4867 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459800 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459820 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459831 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.459843 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd9wz\" (UniqueName: \"kubernetes.io/projected/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba-kube-api-access-pd9wz\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.464711 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config" (OuterVolumeSpecName: "config") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.466453 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb" (OuterVolumeSpecName: "kube-api-access-d5pbb") pod "479f0ca7-1408-41d4-abaf-6034d0802100" (UID: "479f0ca7-1408-41d4-abaf-6034d0802100"). InnerVolumeSpecName "kube-api-access-d5pbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.467343 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf" (OuterVolumeSpecName: "kube-api-access-6n5tf") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "kube-api-access-6n5tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.468338 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.473943 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.475307 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts" (OuterVolumeSpecName: "scripts") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.484352 4867 generic.go:334] "Generic (PLEG): container finished" podID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerID="82e68b2b9f69ed30e1eee05edc30a399c219b479de169d6d7373d0b615332b11" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.484412 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerDied","Data":"82e68b2b9f69ed30e1eee05edc30a399c219b479de169d6d7373d0b615332b11"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.487715 4867 generic.go:334] "Generic (PLEG): container finished" podID="b51b1751-7b06-4118-9eff-961ef320bf22" containerID="f7cd3a86953862b945327d07440fe89481f1546ab42431b9a9517408da200482" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.487818 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerDied","Data":"f7cd3a86953862b945327d07440fe89481f1546ab42431b9a9517408da200482"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.492764 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "479f0ca7-1408-41d4-abaf-6034d0802100" (UID: "479f0ca7-1408-41d4-abaf-6034d0802100"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.493049 4867 generic.go:334] "Generic (PLEG): container finished" podID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerID="e77048aeb4e616861c5d278c417e28346695e6edbd2cfceba516f723cc72bdf9" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.493239 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerDied","Data":"e77048aeb4e616861c5d278c417e28346695e6edbd2cfceba516f723cc72bdf9"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.502027 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerStarted","Data":"35dfd871e3575c594368d88c9965b5034e1b0024ed10fec18e5ec7de472a25cf"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.505561 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.507791 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "479f0ca7-1408-41d4-abaf-6034d0802100" (UID: "479f0ca7-1408-41d4-abaf-6034d0802100"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.518028 4867 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 12 07:12:34 crc kubenswrapper[4867]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-12T07:12:29Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 12 07:12:34 crc kubenswrapper[4867]: /etc/init.d/functions: line 589: 498 Alarm clock "$@" Dec 12 07:12:34 crc kubenswrapper[4867]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-8njtj" message=< Dec 12 07:12:34 crc kubenswrapper[4867]: Exiting ovn-controller (1) [FAILED] Dec 12 07:12:34 crc kubenswrapper[4867]: Killing ovn-controller (1) [ OK ] Dec 12 07:12:34 crc kubenswrapper[4867]: 2025-12-12T07:12:29Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 12 07:12:34 crc kubenswrapper[4867]: /etc/init.d/functions: line 589: 498 Alarm clock "$@" Dec 12 07:12:34 crc kubenswrapper[4867]: > Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.518112 4867 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 12 07:12:34 crc kubenswrapper[4867]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-12T07:12:29Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 12 07:12:34 crc kubenswrapper[4867]: /etc/init.d/functions: line 589: 498 Alarm clock "$@" Dec 12 07:12:34 crc kubenswrapper[4867]: > pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" containerID="cri-o://43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.518165 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" containerID="cri-o://43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" gracePeriod=27 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.531114 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.531253 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.531418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "479f0ca7-1408-41d4-abaf-6034d0802100" (UID: "479f0ca7-1408-41d4-abaf-6034d0802100"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.539086 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron29f2-account-delete-77p2l" podStartSLOduration=4.539067365 podStartE2EDuration="4.539067365s" podCreationTimestamp="2025-12-12 07:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.517471258 +0000 UTC m=+1439.088852527" watchObservedRunningTime="2025-12-12 07:12:31.539067365 +0000 UTC m=+1439.110448634" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.540532 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerStarted","Data":"bddc74aa41be0e6d0235cd6a19049d82e8ea9dad8634a1323dc3669d9156fb82"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.549476 4867 generic.go:334] "Generic (PLEG): container finished" podID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerID="dc60c646092d5fd6e48768335070dbf4389a632716a9616ea70be7334134e290" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.549507 4867 generic.go:334] "Generic (PLEG): container finished" podID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerID="bc30a378f2f47b54d0e25ab70eab05c43f7eb5652e82cd1c2e5b19b297703f21" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.549602 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerDied","Data":"dc60c646092d5fd6e48768335070dbf4389a632716a9616ea70be7334134e290"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.549664 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerDied","Data":"bc30a378f2f47b54d0e25ab70eab05c43f7eb5652e82cd1c2e5b19b297703f21"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.554780 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-92rzj_c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba/openstack-network-exporter/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.554946 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-92rzj" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.555005 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-92rzj" event={"ID":"c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba","Type":"ContainerDied","Data":"416dd3385a5b82f1c0ed7903b247071a9ab944f39f0054abc1ac5eb8c8849739"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.555045 4867 scope.go:117] "RemoveContainer" containerID="c4b46d15a056004b20724580a9403bb36eec9ea66794fe8c2015a7db598e5ae8" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.561934 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.561958 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.561977 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.561988 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.561998 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5pbb\" (UniqueName: \"kubernetes.io/projected/479f0ca7-1408-41d4-abaf-6034d0802100-kube-api-access-d5pbb\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.562009 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.562020 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.562029 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/479f0ca7-1408-41d4-abaf-6034d0802100-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.562037 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n5tf\" (UniqueName: \"kubernetes.io/projected/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-kube-api-access-6n5tf\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.562045 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.565518 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placementc9b0-account-delete-5g27c" podStartSLOduration=5.565498325 podStartE2EDuration="5.565498325s" podCreationTimestamp="2025-12-12 07:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.559828771 +0000 UTC m=+1439.131210040" watchObservedRunningTime="2025-12-12 07:12:31.565498325 +0000 UTC m=+1439.136879594" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.568930 4867 generic.go:334] "Generic (PLEG): container finished" podID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerID="7af7ea533c4fbeeee8dc600c2255f6879f5ae8111689a0c44bb13bbf800546fe" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.569042 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerDied","Data":"7af7ea533c4fbeeee8dc600c2255f6879f5ae8111689a0c44bb13bbf800546fe"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.571932 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0896e-account-delete-9tg4b" event={"ID":"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472","Type":"ContainerStarted","Data":"bfc9805962f8723141ecdc80d5217d80c1b2a7df40ce6757500412eee9680cce"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.577704 4867 generic.go:334] "Generic (PLEG): container finished" podID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerID="14a5c4be4116f44371426de020f2bddb11420dd71b7ecd0f938dff04595d0193" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.577863 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerDied","Data":"14a5c4be4116f44371426de020f2bddb11420dd71b7ecd0f938dff04595d0193"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.586626 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.588185 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.617320 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell0896e-account-delete-9tg4b" podStartSLOduration=4.617284297 podStartE2EDuration="4.617284297s" podCreationTimestamp="2025-12-12 07:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.584727442 +0000 UTC m=+1439.156108721" watchObservedRunningTime="2025-12-12 07:12:31.617284297 +0000 UTC m=+1439.188665696" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.618144 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" (UID: "04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.621178 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerDied","Data":"1499f0119bb5be4db9007f30932c9a39e2288f464937e17ccb75934aea9e9680"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.621176 4867 generic.go:334] "Generic (PLEG): container finished" podID="71501691-562e-4385-a19e-fe9f39cfde61" containerID="1499f0119bb5be4db9007f30932c9a39e2288f464937e17ccb75934aea9e9680" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.622762 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.630179 4867 generic.go:334] "Generic (PLEG): container finished" podID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerID="327346d748281830108225a6a9e113b8722507f57f85ac4bbb3af491d43078a5" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.630280 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf84c603-ce1e-42be-a595-b27ddc77880c","Type":"ContainerDied","Data":"327346d748281830108225a6a9e113b8722507f57f85ac4bbb3af491d43078a5"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.631098 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.632451 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerStarted","Data":"e2a9d914876172e6214a4c0a877ab3a669074c14c1443e7e806e50e7350ff86a"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.637355 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f/ovsdbserver-sb/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.637466 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f","Type":"ContainerDied","Data":"5807753ddb1090827c604ebc45d5015de14ea2b2c68ee8978aefc0d37f32810a"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.637487 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.640574 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-92rzj"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.641337 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerStarted","Data":"9d9541a84d66b2b3172e3884cea7f33b9e6e2538ec678cd2ee81799975bc5831"} Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.642097 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.647130 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.649342 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican74a1-account-delete-m9cms" podStartSLOduration=5.649322579 podStartE2EDuration="5.649322579s" podCreationTimestamp="2025-12-12 07:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.646905568 +0000 UTC m=+1439.218286857" watchObservedRunningTime="2025-12-12 07:12:31.649322579 +0000 UTC m=+1439.220703848" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.651965 4867 generic.go:334] "Generic (PLEG): container finished" podID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerID="7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.652034 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerDied","Data":"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.664769 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.664796 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.664810 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.669790 4867 generic.go:334] "Generic (PLEG): container finished" podID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.669861 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerDied","Data":"969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.670049 4867 scope.go:117] "RemoveContainer" containerID="48b3ddfeb9f53136e00be49e14a43f05d1fd34eb01693af6aa1fd779aa73a984" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.674495 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:31.674555 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="galera" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.675883 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapida12-account-delete-nl9mn" event={"ID":"3c8e566d-e439-4e4b-bb80-759e9c4d7f89","Type":"ContainerStarted","Data":"585c499075ad45a4e241eb7bb0c08e8672b5782b99a85d27b9535747697acf1b"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.682663 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerStarted","Data":"5b52059c4a42740f577bd5d551a5bb09a1e4dd0ae4d8ffaf36f4ba04f2676760"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.689272 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glancea6d9-account-delete-zqpf9" podStartSLOduration=4.68924186 podStartE2EDuration="4.68924186s" podCreationTimestamp="2025-12-12 07:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.675549024 +0000 UTC m=+1439.246930293" watchObservedRunningTime="2025-12-12 07:12:31.68924186 +0000 UTC m=+1439.260623139" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.708391 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.711601 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.742588 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapida12-account-delete-nl9mn" podStartSLOduration=4.742566222 podStartE2EDuration="4.742566222s" podCreationTimestamp="2025-12-12 07:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.718186704 +0000 UTC m=+1439.289567973" watchObservedRunningTime="2025-12-12 07:12:31.742566222 +0000 UTC m=+1439.313947491" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.743449 4867 scope.go:117] "RemoveContainer" containerID="ef733b2c26cb5cf662dfa06f03ae9d9a2ad1b9bfc775896fa039f832b4e8f2b0" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.769757 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinderb8be-account-delete-kvnwc" podStartSLOduration=4.769732491 podStartE2EDuration="4.769732491s" podCreationTimestamp="2025-12-12 07:12:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 07:12:31.737946185 +0000 UTC m=+1439.309327454" watchObservedRunningTime="2025-12-12 07:12:31.769732491 +0000 UTC m=+1439.341113760" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.797657 4867 scope.go:117] "RemoveContainer" containerID="a2c90124c4002721d7428681b4e3e08ef9aee39c1095d4b4a02db73045d4f0df" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.805481 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6796c56757-d2snl" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.162:9696/\": dial tcp 10.217.0.162:9696: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.995844 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.151:9311/healthcheck\": read tcp 10.217.0.2:40358->10.217.0.151:9311: read: connection reset by peer" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.995859 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.151:9311/healthcheck\": read tcp 10.217.0.2:40362->10.217.0.151:9311: read: connection reset by peer" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.998837 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:59990->10.217.0.202:8775: read: connection reset by peer" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:31.998961 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:60002->10.217.0.202:8775: read: connection reset by peer" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.148002 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": read tcp 10.217.0.2:57566->10.217.0.161:8776: read: connection reset by peer" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.176503 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.176578 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data podName:00532992-54f5-4a24-b0e4-eb5512e08753 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.176563801 +0000 UTC m=+1443.747945070 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data") pod "rabbitmq-server-0" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753") : configmap "rabbitmq-config-data" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.274778 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17 is running failed: container process not found" containerID="43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.275215 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17 is running failed: container process not found" containerID="43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.275879 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17 is running failed: container process not found" containerID="43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.275909 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-8njtj" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.292582 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.293037 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.293605 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.293635 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.293835 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.296817 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.297955 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.298027 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.429407 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.430070 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-central-agent" containerID="cri-o://b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.430430 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="proxy-httpd" containerID="cri-o://eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.430477 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-notification-agent" containerID="cri-o://363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.430642 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="sg-core" containerID="cri-o://6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.454392 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.454608 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="437a0623-5c6b-4a29-8da6-f23381fce0ec" containerName="kube-state-metrics" containerID="cri-o://a697abf1dfb4a8d1fd01effec382935f85b0f50a5309476015d25b7269a5f56e" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.636085 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b5bbr"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.655727 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-n2xp2"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.665288 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b5bbr"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.675193 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-n2xp2"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.694645 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.695104 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-74796bf6db-754rv" podUID="38431807-5fbd-4f45-8eca-b42af2c0d221" containerName="keystone-api" containerID="cri-o://a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.713746 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.721172 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8njtj_9a11a089-0048-4c70-94e5-9ae18eadfed5/ovn-controller/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.721216 4867 generic.go:334] "Generic (PLEG): container finished" podID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerID="43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" exitCode=143 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.721308 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj" event={"ID":"9a11a089-0048-4c70-94e5-9ae18eadfed5","Type":"ContainerDied","Data":"43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.724476 4867 generic.go:334] "Generic (PLEG): container finished" podID="765da323-2d60-4494-97bc-f699d9bf8661" containerID="a6b387c7b3b91393496b3df6c1f76038666dc95195810690a8e0c383d5e2aa2d" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.724511 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerDied","Data":"a6b387c7b3b91393496b3df6c1f76038666dc95195810690a8e0c383d5e2aa2d"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.726141 4867 generic.go:334] "Generic (PLEG): container finished" podID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerID="8cd180ed7cb6b465bdfd0105a35071f5be7a9810e30b308406669748fe930aab" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.726181 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerDied","Data":"8cd180ed7cb6b465bdfd0105a35071f5be7a9810e30b308406669748fe930aab"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.730327 4867 generic.go:334] "Generic (PLEG): container finished" podID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerID="17557e0767d1d09b0c76be9d8ba0d9223b9b9565bc8ce63ba2fb3d24cdeacbe1" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.730408 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerDied","Data":"17557e0767d1d09b0c76be9d8ba0d9223b9b9565bc8ce63ba2fb3d24cdeacbe1"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.731587 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-87645"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.738475 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-87645"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.745420 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8cda-account-create-update-hr2p8"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.762903 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8cda-account-create-update-hr2p8"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.764184 4867 generic.go:334] "Generic (PLEG): container finished" podID="163cec12-d32c-4be9-aee0-55703387f2db" containerID="384bb1fccf6418383d28e93ad17cb3f3d26be31108864c9e588482c7982d8d1c" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.764315 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerDied","Data":"384bb1fccf6418383d28e93ad17cb3f3d26be31108864c9e588482c7982d8d1c"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.770443 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-rchnq"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.772278 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican74a1-account-delete-m9cms" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773091 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell0896e-account-delete-9tg4b" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773213 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placementc9b0-account-delete-5g27c" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773581 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinderb8be-account-delete-kvnwc" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773836 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapida12-account-delete-nl9mn" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773593 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron29f2-account-delete-77p2l" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.773880 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancea6d9-account-delete-zqpf9" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.780922 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-rchnq"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.796362 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-da12-account-create-update-wttp8"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.801886 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-da12-account-create-update-wttp8"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.811056 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.825038 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tc5wq"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.859293 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.859415 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" path="/var/lib/kubelet/pods/04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.860823 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.861077 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9cd48c-04e0-4b85-9809-e3c89fbb5081" path="/var/lib/kubelet/pods/0c9cd48c-04e0-4b85-9809-e3c89fbb5081/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.861739 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf99705-8647-4433-bb42-6869440a63c6" path="/var/lib/kubelet/pods/2cf99705-8647-4433-bb42-6869440a63c6/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.862797 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b60ab40-1d50-4c2d-a9f3-726f0825c58a" path="/var/lib/kubelet/pods/3b60ab40-1d50-4c2d-a9f3-726f0825c58a/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.863353 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="479f0ca7-1408-41d4-abaf-6034d0802100" path="/var/lib/kubelet/pods/479f0ca7-1408-41d4-abaf-6034d0802100/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.863896 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f68187-8afa-443c-bce0-1d5aff2a234e" path="/var/lib/kubelet/pods/83f68187-8afa-443c-bce0-1d5aff2a234e/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.864941 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6b7070b-03c8-4d2b-b72a-6474d1e30cc2" path="/var/lib/kubelet/pods/b6b7070b-03c8-4d2b-b72a-6474d1e30cc2/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.865536 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" path="/var/lib/kubelet/pods/c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.866286 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d930b18c-8189-43e2-b59b-9b75e0ac338b" path="/var/lib/kubelet/pods/d930b18c-8189-43e2-b59b-9b75e0ac338b/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.866877 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tc5wq"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.868303 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.868348 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="746ca9c4-7621-489c-86f2-628a9630039c" containerName="nova-scheduler-scheduler" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.871829 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-896e-account-create-update-x9tvp"] Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892197 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892280 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts podName:94f74241-b4d4-41fb-b95f-57534e337198 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.3922609 +0000 UTC m=+1440.963642249 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts") pod "neutron29f2-account-delete-77p2l" (UID: "94f74241-b4d4-41fb-b95f-57534e337198") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892338 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892419 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts podName:cd36e355-181a-4042-a049-aeda5ec0770f nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.392399174 +0000 UTC m=+1440.963780543 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts") pod "barbican74a1-account-delete-m9cms" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892426 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.892452 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts podName:b500049d-056a-40e5-bce8-b6b9f3196d64 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.392444555 +0000 UTC m=+1440.963825934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts") pod "glancea6d9-account-delete-zqpf9" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893461 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893486 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893497 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts podName:66d1496c-b4da-404b-9f9c-da2c405070da nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.39348622 +0000 UTC m=+1440.964867569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts") pod "placementc9b0-account-delete-5g27c" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893515 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts podName:b1b2d98f-1a49-4d66-89b3-ca246c9f227c nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.393507131 +0000 UTC m=+1440.964888480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts") pod "cinderb8be-account-delete-kvnwc" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893542 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893568 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts podName:c7dc7ba4-cc77-4569-ab0f-4bff75ab6472 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.393562843 +0000 UTC m=+1440.964944242 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts") pod "novacell0896e-account-delete-9tg4b" (UID: "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893592 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:32.893609 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts podName:3c8e566d-e439-4e4b-bb80-759e9c4d7f89 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:33.393603784 +0000 UTC m=+1440.964985163 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts") pod "novaapida12-account-delete-nl9mn" (UID: "3c8e566d-e439-4e4b-bb80-759e9c4d7f89") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.895583 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.902191 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-896e-account-create-update-x9tvp"] Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:32.944606 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="galera" containerID="cri-o://456526542b9c5797b244ddfdcbdceb1b0f47fb01b1e916fbb187836108af239b" gracePeriod=30 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.201745 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-67c5f5d445-rd5dx" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.164:8080/healthcheck\": dial tcp 10.217.0.164:8080: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.201748 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-67c5f5d445-rd5dx" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.164:8080/healthcheck\": dial tcp 10.217.0.164:8080: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405297 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405370 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405407 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405483 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405540 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405546 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405499 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts podName:cd36e355-181a-4042-a049-aeda5ec0770f nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.405476197 +0000 UTC m=+1441.976857536 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts") pod "barbican74a1-account-delete-m9cms" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405602 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts podName:94f74241-b4d4-41fb-b95f-57534e337198 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.405592469 +0000 UTC m=+1441.976973738 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts") pod "neutron29f2-account-delete-77p2l" (UID: "94f74241-b4d4-41fb-b95f-57534e337198") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405615 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts podName:b500049d-056a-40e5-bce8-b6b9f3196d64 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.4056085 +0000 UTC m=+1441.976989769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts") pod "glancea6d9-account-delete-zqpf9" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405626 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts podName:c7dc7ba4-cc77-4569-ab0f-4bff75ab6472 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.40562059 +0000 UTC m=+1441.977001969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts") pod "novacell0896e-account-delete-9tg4b" (UID: "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405629 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405647 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts podName:b1b2d98f-1a49-4d66-89b3-ca246c9f227c nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.405640251 +0000 UTC m=+1441.977021520 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts") pod "cinderb8be-account-delete-kvnwc" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405665 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts podName:3c8e566d-e439-4e4b-bb80-759e9c4d7f89 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.405658411 +0000 UTC m=+1441.977039680 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts") pod "novaapida12-account-delete-nl9mn" (UID: "3c8e566d-e439-4e4b-bb80-759e9c4d7f89") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:33.405678 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts podName:66d1496c-b4da-404b-9f9c-da2c405070da nodeName:}" failed. No retries permitted until 2025-12-12 07:12:34.405671581 +0000 UTC m=+1441.977052850 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts") pod "placementc9b0-account-delete-5g27c" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.737970 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.195:6080/vnc_lite.html\": dial tcp 10.217.0.195:6080: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.781564 4867 generic.go:334] "Generic (PLEG): container finished" podID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerID="5b52059c4a42740f577bd5d551a5bb09a1e4dd0ae4d8ffaf36f4ba04f2676760" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.781584 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerDied","Data":"5b52059c4a42740f577bd5d551a5bb09a1e4dd0ae4d8ffaf36f4ba04f2676760"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.782385 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinderb8be-account-delete-kvnwc" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.782448 4867 scope.go:117] "RemoveContainer" containerID="5b52059c4a42740f577bd5d551a5bb09a1e4dd0ae4d8ffaf36f4ba04f2676760" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787624 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerID="eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787649 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerID="6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474" exitCode=2 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787660 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerID="b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787707 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerDied","Data":"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787747 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerDied","Data":"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.787760 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerDied","Data":"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.800115 4867 generic.go:334] "Generic (PLEG): container finished" podID="66d1496c-b4da-404b-9f9c-da2c405070da" containerID="bddc74aa41be0e6d0235cd6a19049d82e8ea9dad8634a1323dc3669d9156fb82" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.800202 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerDied","Data":"bddc74aa41be0e6d0235cd6a19049d82e8ea9dad8634a1323dc3669d9156fb82"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.801147 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placementc9b0-account-delete-5g27c" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.801189 4867 scope.go:117] "RemoveContainer" containerID="bddc74aa41be0e6d0235cd6a19049d82e8ea9dad8634a1323dc3669d9156fb82" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.801969 4867 generic.go:334] "Generic (PLEG): container finished" podID="437a0623-5c6b-4a29-8da6-f23381fce0ec" containerID="a697abf1dfb4a8d1fd01effec382935f85b0f50a5309476015d25b7269a5f56e" exitCode=2 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.802003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"437a0623-5c6b-4a29-8da6-f23381fce0ec","Type":"ContainerDied","Data":"a697abf1dfb4a8d1fd01effec382935f85b0f50a5309476015d25b7269a5f56e"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.804709 4867 generic.go:334] "Generic (PLEG): container finished" podID="3c8e566d-e439-4e4b-bb80-759e9c4d7f89" containerID="585c499075ad45a4e241eb7bb0c08e8672b5782b99a85d27b9535747697acf1b" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.804802 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapida12-account-delete-nl9mn" event={"ID":"3c8e566d-e439-4e4b-bb80-759e9c4d7f89","Type":"ContainerDied","Data":"585c499075ad45a4e241eb7bb0c08e8672b5782b99a85d27b9535747697acf1b"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.807707 4867 generic.go:334] "Generic (PLEG): container finished" podID="94f74241-b4d4-41fb-b95f-57534e337198" containerID="35dfd871e3575c594368d88c9965b5034e1b0024ed10fec18e5ec7de472a25cf" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.807749 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerDied","Data":"35dfd871e3575c594368d88c9965b5034e1b0024ed10fec18e5ec7de472a25cf"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.808539 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron29f2-account-delete-77p2l" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.808590 4867 scope.go:117] "RemoveContainer" containerID="35dfd871e3575c594368d88c9965b5034e1b0024ed10fec18e5ec7de472a25cf" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.810899 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00269882-fe60-4b26-8d5c-1b34b014b191/ovn-northd/0.log" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.810929 4867 generic.go:334] "Generic (PLEG): container finished" podID="00269882-fe60-4b26-8d5c-1b34b014b191" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" exitCode=139 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.810975 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerDied","Data":"7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.819256 4867 generic.go:334] "Generic (PLEG): container finished" podID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerID="9d9541a84d66b2b3172e3884cea7f33b9e6e2538ec678cd2ee81799975bc5831" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.819309 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerDied","Data":"9d9541a84d66b2b3172e3884cea7f33b9e6e2538ec678cd2ee81799975bc5831"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.819940 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancea6d9-account-delete-zqpf9" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.819975 4867 scope.go:117] "RemoveContainer" containerID="9d9541a84d66b2b3172e3884cea7f33b9e6e2538ec678cd2ee81799975bc5831" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.822926 4867 generic.go:334] "Generic (PLEG): container finished" podID="cd36e355-181a-4042-a049-aeda5ec0770f" containerID="e2a9d914876172e6214a4c0a877ab3a669074c14c1443e7e806e50e7350ff86a" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.822961 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerDied","Data":"e2a9d914876172e6214a4c0a877ab3a669074c14c1443e7e806e50e7350ff86a"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.823245 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican74a1-account-delete-m9cms" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.823264 4867 scope.go:117] "RemoveContainer" containerID="e2a9d914876172e6214a4c0a877ab3a669074c14c1443e7e806e50e7350ff86a" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.828779 4867 generic.go:334] "Generic (PLEG): container finished" podID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerID="ee6e5dae4c9a11595f4e81ee6c481fdc2279b8c9a6898d55722fd9690e003f40" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.828867 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerDied","Data":"ee6e5dae4c9a11595f4e81ee6c481fdc2279b8c9a6898d55722fd9690e003f40"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.833515 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" containerID="bfc9805962f8723141ecdc80d5217d80c1b2a7df40ce6757500412eee9680cce" exitCode=1 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.833606 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0896e-account-delete-9tg4b" event={"ID":"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472","Type":"ContainerDied","Data":"bfc9805962f8723141ecdc80d5217d80c1b2a7df40ce6757500412eee9680cce"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.841577 4867 generic.go:334] "Generic (PLEG): container finished" podID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerID="4a8ad42ecdb710efacb437807e79af57dea08fcccf32bfdc46ca2fca80f02ba9" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.841658 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerDied","Data":"4a8ad42ecdb710efacb437807e79af57dea08fcccf32bfdc46ca2fca80f02ba9"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.858449 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": dial tcp 10.217.0.202:8775: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.858599 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": dial tcp 10.217.0.202:8775: connect: connection refused" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.861632 4867 generic.go:334] "Generic (PLEG): container finished" podID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerID="f164d088b657a841b64a09bed96dee589e4f7130244d17533d2c9a7d6fdb6dab" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:33.861669 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerDied","Data":"f164d088b657a841b64a09bed96dee589e4f7130244d17533d2c9a7d6fdb6dab"} Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429385 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429792 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts podName:c7dc7ba4-cc77-4569-ab0f-4bff75ab6472 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.429752776 +0000 UTC m=+1444.001134045 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts") pod "novacell0896e-account-delete-9tg4b" (UID: "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429394 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429424 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429447 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429510 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429531 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429849 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts podName:3c8e566d-e439-4e4b-bb80-759e9c4d7f89 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.429836228 +0000 UTC m=+1444.001217497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts") pod "novaapida12-account-delete-nl9mn" (UID: "3c8e566d-e439-4e4b-bb80-759e9c4d7f89") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.429997 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts podName:94f74241-b4d4-41fb-b95f-57534e337198 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.429977211 +0000 UTC m=+1444.001358480 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts") pod "neutron29f2-account-delete-77p2l" (UID: "94f74241-b4d4-41fb-b95f-57534e337198") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.430020 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts podName:66d1496c-b4da-404b-9f9c-da2c405070da nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.430014162 +0000 UTC m=+1444.001395431 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts") pod "placementc9b0-account-delete-5g27c" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.430034 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts podName:b1b2d98f-1a49-4d66-89b3-ca246c9f227c nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.430029312 +0000 UTC m=+1444.001410571 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts") pod "cinderb8be-account-delete-kvnwc" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.430051 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts podName:cd36e355-181a-4042-a049-aeda5ec0770f nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.430044813 +0000 UTC m=+1444.001426082 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts") pod "barbican74a1-account-delete-m9cms" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.430361 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.430410 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts podName:b500049d-056a-40e5-bce8-b6b9f3196d64 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:36.430401031 +0000 UTC m=+1444.001782380 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts") pod "glancea6d9-account-delete-zqpf9" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64") : configmap "openstack-scripts" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.633407 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:34 crc kubenswrapper[4867]: E1212 07:12:34.633486 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data podName:65ad43b4-b2fa-4089-b471-66fbae81fbf2 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:42.633468608 +0000 UTC m=+1450.204849867 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data") pod "rabbitmq-cell1-server-0" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2") : configmap "rabbitmq-cell1-config-data" not found Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.855301 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fced4c1-5bda-4c55-bcf9-80c622893f7d" path="/var/lib/kubelet/pods/6fced4c1-5bda-4c55-bcf9-80c622893f7d/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.855871 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6e178c2-60a2-40b1-864d-86fafb2b4d57" path="/var/lib/kubelet/pods/c6e178c2-60a2-40b1-864d-86fafb2b4d57/volumes" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.882398 4867 generic.go:334] "Generic (PLEG): container finished" podID="746ca9c4-7621-489c-86f2-628a9630039c" containerID="aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" exitCode=0 Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.882485 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"746ca9c4-7621-489c-86f2-628a9630039c","Type":"ContainerDied","Data":"aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.884545 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerStarted","Data":"db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.885107 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placementc9b0-account-delete-5g27c" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.888750 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerStarted","Data":"1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.889203 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican74a1-account-delete-m9cms" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.893516 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerStarted","Data":"092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.894377 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron29f2-account-delete-77p2l" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.901143 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerStarted","Data":"ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.906070 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancea6d9-account-delete-zqpf9" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.915206 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerStarted","Data":"00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3"} Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.915838 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinderb8be-account-delete-kvnwc" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:34 crc kubenswrapper[4867]: I1212 07:12:34.936910 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.161:8776/healthcheck\": dial tcp 10.217.0.161:8776: connect: connection refused" Dec 12 07:12:35 crc kubenswrapper[4867]: E1212 07:12:35.369769 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844 is running failed: container process not found" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:35 crc kubenswrapper[4867]: E1212 07:12:35.370248 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844 is running failed: container process not found" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:35 crc kubenswrapper[4867]: E1212 07:12:35.370581 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844 is running failed: container process not found" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 12 07:12:35 crc kubenswrapper[4867]: E1212 07:12:35.370610 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.652735 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.659280 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.673676 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.689213 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.689363 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.697501 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.708800 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00269882-fe60-4b26-8d5c-1b34b014b191/ovn-northd/0.log" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.708901 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.709268 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.721271 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.733802 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.736270 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.748709 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8njtj_9a11a089-0048-4c70-94e5-9ae18eadfed5/ovn-controller/0.log" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.762425 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.763540 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.764391 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.769180 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.772370 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.775364 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data\") pod \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.775722 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776088 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776192 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776280 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776344 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776412 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppg6d\" (UniqueName: \"kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776493 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776562 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776627 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776698 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776755 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjt7k\" (UniqueName: \"kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.776832 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.777140 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.777206 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779272 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5fmv\" (UniqueName: \"kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779349 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779414 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztx68\" (UniqueName: \"kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779517 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779761 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779895 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.779957 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780018 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs\") pod \"bf84c603-ce1e-42be-a595-b27ddc77880c\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780131 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs\") pod \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780212 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780362 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780432 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz79s\" (UniqueName: \"kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s\") pod \"746ca9c4-7621-489c-86f2-628a9630039c\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.780499 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781219 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781316 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs\") pod \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781381 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781449 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781515 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781608 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xl4x\" (UniqueName: \"kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x\") pod \"bf84c603-ce1e-42be-a595-b27ddc77880c\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781675 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.781951 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782031 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wsd7\" (UniqueName: \"kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782094 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782156 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782219 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data\") pod \"746ca9c4-7621-489c-86f2-628a9630039c\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782306 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle\") pod \"bf84c603-ce1e-42be-a595-b27ddc77880c\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782380 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782446 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data\") pod \"bf84c603-ce1e-42be-a595-b27ddc77880c\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65crj\" (UniqueName: \"kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782563 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782624 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id\") pod \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\" (UID: \"fcb3e56a-fd7b-4d8e-b42d-2d55de745667\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.782686 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783127 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qph4s\" (UniqueName: \"kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783170 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783206 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data\") pod \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\" (UID: \"7794fa05-fbb0-4c40-9711-60f299e5ab5b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783245 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs\") pod \"765da323-2d60-4494-97bc-f699d9bf8661\" (UID: \"765da323-2d60-4494-97bc-f699d9bf8661\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783263 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783284 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data\") pod \"aa35f883-b75a-4e67-aeed-f36d00a074cf\" (UID: \"aa35f883-b75a-4e67-aeed-f36d00a074cf\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783307 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle\") pod \"746ca9c4-7621-489c-86f2-628a9630039c\" (UID: \"746ca9c4-7621-489c-86f2-628a9630039c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783328 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcfbs\" (UniqueName: \"kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs\") pod \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783370 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2vc9\" (UniqueName: \"kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9\") pod \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783412 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config\") pod \"00269882-fe60-4b26-8d5c-1b34b014b191\" (UID: \"00269882-fe60-4b26-8d5c-1b34b014b191\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783430 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783452 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle\") pod \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\" (UID: \"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783479 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs\") pod \"163cec12-d32c-4be9-aee0-55703387f2db\" (UID: \"163cec12-d32c-4be9-aee0-55703387f2db\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783527 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs\") pod \"bf84c603-ce1e-42be-a595-b27ddc77880c\" (UID: \"bf84c603-ce1e-42be-a595-b27ddc77880c\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783555 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts\") pod \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\" (UID: \"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.783573 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs\") pod \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\" (UID: \"50a1fd31-a115-4bb6-b65e-70adc5dd77ed\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.777199 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.778465 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.778625 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs" (OuterVolumeSpecName: "logs") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.778944 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts" (OuterVolumeSpecName: "scripts") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.787908 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d" (OuterVolumeSpecName: "kube-api-access-ppg6d") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "kube-api-access-ppg6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.794604 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts" (OuterVolumeSpecName: "scripts") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.797536 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config" (OuterVolumeSpecName: "config") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.797657 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.803508 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs" (OuterVolumeSpecName: "logs") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.803725 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv" (OuterVolumeSpecName: "kube-api-access-w5fmv") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "kube-api-access-w5fmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.803759 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.805527 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" (UID: "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.806294 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs" (OuterVolumeSpecName: "logs") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.806954 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts" (OuterVolumeSpecName: "scripts") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.807112 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.808266 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs" (OuterVolumeSpecName: "logs") pod "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" (UID: "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.810928 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs" (OuterVolumeSpecName: "logs") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.811035 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s" (OuterVolumeSpecName: "kube-api-access-jz79s") pod "746ca9c4-7621-489c-86f2-628a9630039c" (UID: "746ca9c4-7621-489c-86f2-628a9630039c"). InnerVolumeSpecName "kube-api-access-jz79s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.811125 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s" (OuterVolumeSpecName: "kube-api-access-qph4s") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "kube-api-access-qph4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.814565 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.814809 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts" (OuterVolumeSpecName: "scripts") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.814996 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs" (OuterVolumeSpecName: "logs") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.815176 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k" (OuterVolumeSpecName: "kube-api-access-sjt7k") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "kube-api-access-sjt7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.815188 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj" (OuterVolumeSpecName: "kube-api-access-65crj") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "kube-api-access-65crj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.815270 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9" (OuterVolumeSpecName: "kube-api-access-s2vc9") pod "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" (UID: "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae"). InnerVolumeSpecName "kube-api-access-s2vc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.815301 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7" (OuterVolumeSpecName: "kube-api-access-6wsd7") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "kube-api-access-6wsd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.815755 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.816747 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x" (OuterVolumeSpecName: "kube-api-access-7xl4x") pod "bf84c603-ce1e-42be-a595-b27ddc77880c" (UID: "bf84c603-ce1e-42be-a595-b27ddc77880c"). InnerVolumeSpecName "kube-api-access-7xl4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.817485 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.817538 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts" (OuterVolumeSpecName: "scripts") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.819119 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68" (OuterVolumeSpecName: "kube-api-access-ztx68") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "kube-api-access-ztx68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.819138 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts" (OuterVolumeSpecName: "scripts") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.825068 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.839415 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs" (OuterVolumeSpecName: "kube-api-access-lcfbs") pod "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" (UID: "c7dc7ba4-cc77-4569-ab0f-4bff75ab6472"). InnerVolumeSpecName "kube-api-access-lcfbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.857432 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.864363 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data" (OuterVolumeSpecName: "config-data") pod "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" (UID: "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.877771 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886531 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886614 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cjs2\" (UniqueName: \"kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2\") pod \"437a0623-5c6b-4a29-8da6-f23381fce0ec\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886654 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config\") pod \"437a0623-5c6b-4a29-8da6-f23381fce0ec\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886678 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run" (OuterVolumeSpecName: "var-run") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886688 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts\") pod \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886715 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qfbb\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886742 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886771 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886800 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886831 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcqbl\" (UniqueName: \"kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886856 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886920 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.886957 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887091 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fg6n\" (UniqueName: \"kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n\") pod \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\" (UID: \"3c8e566d-e439-4e4b-bb80-759e9c4d7f89\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887174 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs\") pod \"437a0623-5c6b-4a29-8da6-f23381fce0ec\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887200 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887322 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts\") pod \"9a11a089-0048-4c70-94e5-9ae18eadfed5\" (UID: \"9a11a089-0048-4c70-94e5-9ae18eadfed5\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887361 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle\") pod \"437a0623-5c6b-4a29-8da6-f23381fce0ec\" (UID: \"437a0623-5c6b-4a29-8da6-f23381fce0ec\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887394 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887429 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.887504 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd\") pod \"56091088-518c-4636-bb5e-4bdcc9f0397b\" (UID: \"56091088-518c-4636-bb5e-4bdcc9f0397b\") " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888383 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888411 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888425 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65crj\" (UniqueName: \"kubernetes.io/projected/00269882-fe60-4b26-8d5c-1b34b014b191-kube-api-access-65crj\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888438 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888460 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888475 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qph4s\" (UniqueName: \"kubernetes.io/projected/765da323-2d60-4494-97bc-f699d9bf8661-kube-api-access-qph4s\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888494 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888507 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/765da323-2d60-4494-97bc-f699d9bf8661-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888520 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcfbs\" (UniqueName: \"kubernetes.io/projected/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-kube-api-access-lcfbs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888533 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888545 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2vc9\" (UniqueName: \"kubernetes.io/projected/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-kube-api-access-s2vc9\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888558 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00269882-fe60-4b26-8d5c-1b34b014b191-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888570 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163cec12-d32c-4be9-aee0-55703387f2db-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888581 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888592 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888602 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888615 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888627 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888637 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888649 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppg6d\" (UniqueName: \"kubernetes.io/projected/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-kube-api-access-ppg6d\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888661 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163cec12-d32c-4be9-aee0-55703387f2db-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888672 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888683 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjt7k\" (UniqueName: \"kubernetes.io/projected/163cec12-d32c-4be9-aee0-55703387f2db-kube-api-access-sjt7k\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888694 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5fmv\" (UniqueName: \"kubernetes.io/projected/7794fa05-fbb0-4c40-9711-60f299e5ab5b-kube-api-access-w5fmv\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888706 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888717 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztx68\" (UniqueName: \"kubernetes.io/projected/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-kube-api-access-ztx68\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888729 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888740 4867 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888750 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888761 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888772 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz79s\" (UniqueName: \"kubernetes.io/projected/746ca9c4-7621-489c-86f2-628a9630039c-kube-api-access-jz79s\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888784 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888794 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa35f883-b75a-4e67-aeed-f36d00a074cf-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888807 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888819 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888831 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7794fa05-fbb0-4c40-9711-60f299e5ab5b-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888842 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xl4x\" (UniqueName: \"kubernetes.io/projected/bf84c603-ce1e-42be-a595-b27ddc77880c-kube-api-access-7xl4x\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.888854 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wsd7\" (UniqueName: \"kubernetes.io/projected/aa35f883-b75a-4e67-aeed-f36d00a074cf-kube-api-access-6wsd7\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.889319 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts" (OuterVolumeSpecName: "scripts") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.889987 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.890550 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.890611 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.892109 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c8e566d-e439-4e4b-bb80-759e9c4d7f89" (UID: "3c8e566d-e439-4e4b-bb80-759e9c4d7f89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.893295 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.893346 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.896103 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb" (OuterVolumeSpecName: "kube-api-access-5qfbb") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "kube-api-access-5qfbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.896891 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl" (OuterVolumeSpecName: "kube-api-access-lcqbl") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "kube-api-access-lcqbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.898349 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2" (OuterVolumeSpecName: "kube-api-access-2cjs2") pod "437a0623-5c6b-4a29-8da6-f23381fce0ec" (UID: "437a0623-5c6b-4a29-8da6-f23381fce0ec"). InnerVolumeSpecName "kube-api-access-2cjs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.909656 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.918413 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n" (OuterVolumeSpecName: "kube-api-access-6fg6n") pod "3c8e566d-e439-4e4b-bb80-759e9c4d7f89" (UID: "3c8e566d-e439-4e4b-bb80-759e9c4d7f89"). InnerVolumeSpecName "kube-api-access-6fg6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.918690 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.941753 4867 generic.go:334] "Generic (PLEG): container finished" podID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerID="49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" exitCode=0 Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.941824 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerDied","Data":"49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e"} Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.956244 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf84c603-ce1e-42be-a595-b27ddc77880c" (UID: "bf84c603-ce1e-42be-a595-b27ddc77880c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.968950 4867 generic.go:334] "Generic (PLEG): container finished" podID="66d1496c-b4da-404b-9f9c-da2c405070da" containerID="db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e" exitCode=1 Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.969016 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerDied","Data":"db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e"} Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.969076 4867 scope.go:117] "RemoveContainer" containerID="bddc74aa41be0e6d0235cd6a19049d82e8ea9dad8634a1323dc3669d9156fb82" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.969640 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placementc9b0-account-delete-5g27c" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.969678 4867 scope.go:117] "RemoveContainer" containerID="db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e" Dec 12 07:12:35 crc kubenswrapper[4867]: E1212 07:12:35.969908 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=placementc9b0-account-delete-5g27c_openstack(66d1496c-b4da-404b-9f9c-da2c405070da)\"" pod="openstack/placementc9b0-account-delete-5g27c" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.979083 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.979050 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"437a0623-5c6b-4a29-8da6-f23381fce0ec","Type":"ContainerDied","Data":"d03b8a8f951c2b034df4288e68257eee6aeac9891b694168ecfaded3896fa122"} Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.989767 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6f69644fbd-dsggh" event={"ID":"7794fa05-fbb0-4c40-9711-60f299e5ab5b","Type":"ContainerDied","Data":"cb590d6285be23ac892b56c5aef2863e103571d677ab794adca39f7fb7e606fc"} Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990474 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990502 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990514 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fg6n\" (UniqueName: \"kubernetes.io/projected/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-kube-api-access-6fg6n\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990525 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a11a089-0048-4c70-94e5-9ae18eadfed5-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990533 4867 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990542 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990550 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56091088-518c-4636-bb5e-4bdcc9f0397b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990559 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990568 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cjs2\" (UniqueName: \"kubernetes.io/projected/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-api-access-2cjs2\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990576 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8e566d-e439-4e4b-bb80-759e9c4d7f89-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990584 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qfbb\" (UniqueName: \"kubernetes.io/projected/56091088-518c-4636-bb5e-4bdcc9f0397b-kube-api-access-5qfbb\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990593 4867 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990601 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcqbl\" (UniqueName: \"kubernetes.io/projected/9a11a089-0048-4c70-94e5-9ae18eadfed5-kube-api-access-lcqbl\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.990609 4867 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a11a089-0048-4c70-94e5-9ae18eadfed5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.991392 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6f69644fbd-dsggh" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.992901 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0896e-account-delete-9tg4b" event={"ID":"c7dc7ba4-cc77-4569-ab0f-4bff75ab6472","Type":"ContainerDied","Data":"428e5cb15516a50e79f56f0b9879e83f4c6464aa13842b57b5ba94dfa10f28d1"} Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.992963 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0896e-account-delete-9tg4b" Dec 12 07:12:35 crc kubenswrapper[4867]: I1212 07:12:35.997147 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"163cec12-d32c-4be9-aee0-55703387f2db","Type":"ContainerDied","Data":"8e9487be27e4f37fb16b1e5172bed50b567bef99991375dcb870b256497c8730"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:35.999328 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.008920 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"aa35f883-b75a-4e67-aeed-f36d00a074cf","Type":"ContainerDied","Data":"688b563243ecf18f051dc395d4e9bab540550c17791a5cf14397013544eebe33"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.009137 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.021026 4867 generic.go:334] "Generic (PLEG): container finished" podID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerID="ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f" exitCode=1 Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.021152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerDied","Data":"ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.029341 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glancea6d9-account-delete-zqpf9" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.029395 4867 scope.go:117] "RemoveContainer" containerID="ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.029995 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=glancea6d9-account-delete-zqpf9_openstack(b500049d-056a-40e5-bce8-b6b9f3196d64)\"" pod="openstack/glancea6d9-account-delete-zqpf9" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.034123 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf84c603-ce1e-42be-a595-b27ddc77880c","Type":"ContainerDied","Data":"fe6d5d260958339042003dfb25398c1fed02916feffc9dd34e55df3164025009"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.034324 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.041090 4867 generic.go:334] "Generic (PLEG): container finished" podID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerID="00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3" exitCode=1 Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.041162 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerDied","Data":"00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.046405 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinderb8be-account-delete-kvnwc" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.046449 4867 scope.go:117] "RemoveContainer" containerID="00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.046651 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=cinderb8be-account-delete-kvnwc_openstack(b1b2d98f-1a49-4d66-89b3-ca246c9f227c)\"" pod="openstack/cinderb8be-account-delete-kvnwc" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.046792 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.047618 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fcb3e56a-fd7b-4d8e-b42d-2d55de745667","Type":"ContainerDied","Data":"becc0616fc52a7219276f1a0c33ac193b68cf392b6995c852033103fcc75c52e"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.047726 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.055550 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapida12-account-delete-nl9mn" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.055542 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapida12-account-delete-nl9mn" event={"ID":"3c8e566d-e439-4e4b-bb80-759e9c4d7f89","Type":"ContainerDied","Data":"95bcf6dd988818d74803c0b4f049b2ce430384f2bf06f2d6c67e3c6a960c0d61"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.058142 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data" (OuterVolumeSpecName: "config-data") pod "bf84c603-ce1e-42be-a595-b27ddc77880c" (UID: "bf84c603-ce1e-42be-a595-b27ddc77880c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.058633 4867 generic.go:334] "Generic (PLEG): container finished" podID="b51b1751-7b06-4118-9eff-961ef320bf22" containerID="b6413fee15f4b9393e64a3a1314ffb75d76cc3621c445d54a57ddbb7d9e78327" exitCode=0 Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.058690 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerDied","Data":"b6413fee15f4b9393e64a3a1314ffb75d76cc3621c445d54a57ddbb7d9e78327"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.060567 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8njtj_9a11a089-0048-4c70-94e5-9ae18eadfed5/ovn-controller/0.log" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.060678 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8njtj" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.060706 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8njtj" event={"ID":"9a11a089-0048-4c70-94e5-9ae18eadfed5","Type":"ContainerDied","Data":"aebb55f232f65de131c725eada57789830eee1d462d92db26d07c3d12f0c96f3"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.062589 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.062772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"746ca9c4-7621-489c-86f2-628a9630039c","Type":"ContainerDied","Data":"abbaa3c5bc249a78813a008ddc5a0d13a0c2722e39516639596ad3fbaa74c3e6"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.073432 4867 generic.go:334] "Generic (PLEG): container finished" podID="cd36e355-181a-4042-a049-aeda5ec0770f" containerID="1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e" exitCode=1 Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.073474 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerDied","Data":"1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.074352 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbican74a1-account-delete-m9cms" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.074391 4867 scope.go:117] "RemoveContainer" containerID="1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.074748 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=barbican74a1-account-delete-m9cms_openstack(cd36e355-181a-4042-a049-aeda5ec0770f)\"" pod="openstack/barbican74a1-account-delete-m9cms" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.080447 4867 scope.go:117] "RemoveContainer" containerID="a697abf1dfb4a8d1fd01effec382935f85b0f50a5309476015d25b7269a5f56e" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.081766 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fbb7b7c4-4c08-4f29-8e06-18af9f7422ae","Type":"ContainerDied","Data":"282dc848c0d4f9b95110e2c42055c000a3d4db533e84c29c0e761f5689491da3"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.081864 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.089182 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_00269882-fe60-4b26-8d5c-1b34b014b191/ovn-northd/0.log" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.089314 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"00269882-fe60-4b26-8d5c-1b34b014b191","Type":"ContainerDied","Data":"f85d8ccfd37c9b440bd05e12368df61b028435fabdd7c690acedf5b5bcbf9501"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.089371 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.092827 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.092860 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.097252 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.097978 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.098022 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"50a1fd31-a115-4bb6-b65e-70adc5dd77ed","Type":"ContainerDied","Data":"753cac05b80813598a54d00c3be296ae0bf594560258313891cd470db8a16304"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.101310 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68ccfcbfb4-q47vm" event={"ID":"765da323-2d60-4494-97bc-f699d9bf8661","Type":"ContainerDied","Data":"1a59081fd9e9b81afc0c2876cdea2a53be58c7cb530b3ea0ca07cb37fdd47e61"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.101343 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68ccfcbfb4-q47vm" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.117461 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-67c5f5d445-rd5dx" event={"ID":"56091088-518c-4636-bb5e-4bdcc9f0397b","Type":"ContainerDied","Data":"b4cea50b4a2409e5c06031d89001030cb96b03f21678fb75ce73c9b0ea30ae49"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.117602 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-67c5f5d445-rd5dx" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.120877 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" (UID: "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.122423 4867 scope.go:117] "RemoveContainer" containerID="17557e0767d1d09b0c76be9d8ba0d9223b9b9565bc8ce63ba2fb3d24cdeacbe1" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.126024 4867 generic.go:334] "Generic (PLEG): container finished" podID="94f74241-b4d4-41fb-b95f-57534e337198" containerID="092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0" exitCode=1 Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.126057 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerDied","Data":"092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0"} Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.126162 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0896e-account-delete-9tg4b"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.126899 4867 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron29f2-account-delete-77p2l" secret="" err="secret \"galera-openstack-dockercfg-lpdhb\" not found" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.126946 4867 scope.go:117] "RemoveContainer" containerID="092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.127294 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=neutron29f2-account-delete-77p2l_openstack(94f74241-b4d4-41fb-b95f-57534e337198)\"" pod="openstack/neutron29f2-account-delete-77p2l" podUID="94f74241-b4d4-41fb-b95f-57534e337198" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.160057 4867 scope.go:117] "RemoveContainer" containerID="e77048aeb4e616861c5d278c417e28346695e6edbd2cfceba516f723cc72bdf9" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.164601 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data" (OuterVolumeSpecName: "config-data") pod "746ca9c4-7621-489c-86f2-628a9630039c" (UID: "746ca9c4-7621-489c-86f2-628a9630039c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.173251 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.182069 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.184157 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.194948 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.195012 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.195027 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.195039 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.195159 4867 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.195365 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data podName:00532992-54f5-4a24-b0e4-eb5512e08753 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:44.195202569 +0000 UTC m=+1451.766583838 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data") pod "rabbitmq-server-0" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753") : configmap "rabbitmq-config-data" not found Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.196607 4867 scope.go:117] "RemoveContainer" containerID="bfc9805962f8723141ecdc80d5217d80c1b2a7df40ce6757500412eee9680cce" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.203178 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapida12-account-delete-nl9mn"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.208705 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.211284 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.228619 4867 scope.go:117] "RemoveContainer" containerID="384bb1fccf6418383d28e93ad17cb3f3d26be31108864c9e588482c7982d8d1c" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.255605 4867 scope.go:117] "RemoveContainer" containerID="f3626d4c1ef6132f0076b910eab89601c61a779670954414124b9e2ed04cba99" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.256781 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data" (OuterVolumeSpecName: "config-data") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.261143 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" (UID: "fbb7b7c4-4c08-4f29-8e06-18af9f7422ae"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.292965 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.293797 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297361 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297387 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297398 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297408 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297418 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.297427 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.301350 4867 scope.go:117] "RemoveContainer" containerID="4a8ad42ecdb710efacb437807e79af57dea08fcccf32bfdc46ca2fca80f02ba9" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.301389 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data" (OuterVolumeSpecName: "config-data") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.307001 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data" (OuterVolumeSpecName: "config-data") pod "aa35f883-b75a-4e67-aeed-f36d00a074cf" (UID: "aa35f883-b75a-4e67-aeed-f36d00a074cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.315348 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "437a0623-5c6b-4a29-8da6-f23381fce0ec" (UID: "437a0623-5c6b-4a29-8da6-f23381fce0ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.339474 4867 scope.go:117] "RemoveContainer" containerID="e0e23c15cb95fadeb8c5fb6992fb31533b6edb8b32c8f954630c0f55ff2e3c91" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.360462 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.365568 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "bf84c603-ce1e-42be-a595-b27ddc77880c" (UID: "bf84c603-ce1e-42be-a595-b27ddc77880c"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.368182 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.399164 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa35f883-b75a-4e67-aeed-f36d00a074cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.399198 4867 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.399213 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.399242 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.418793 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.423480 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.450014 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.454487 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "437a0623-5c6b-4a29-8da6-f23381fce0ec" (UID: "437a0623-5c6b-4a29-8da6-f23381fce0ec"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.456282 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "746ca9c4-7621-489c-86f2-628a9630039c" (UID: "746ca9c4-7621-489c-86f2-628a9630039c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.456670 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.462376 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.467501 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data" (OuterVolumeSpecName: "config-data") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.468179 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "bf84c603-ce1e-42be-a595-b27ddc77880c" (UID: "bf84c603-ce1e-42be-a595-b27ddc77880c"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.471630 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.478339 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.479316 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.480842 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56091088-518c-4636-bb5e-4bdcc9f0397b" (UID: "56091088-518c-4636-bb5e-4bdcc9f0397b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.489504 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data" (OuterVolumeSpecName: "config-data") pod "50a1fd31-a115-4bb6-b65e-70adc5dd77ed" (UID: "50a1fd31-a115-4bb6-b65e-70adc5dd77ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.492810 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data" (OuterVolumeSpecName: "config-data") pod "163cec12-d32c-4be9-aee0-55703387f2db" (UID: "163cec12-d32c-4be9-aee0-55703387f2db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500021 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data" (OuterVolumeSpecName: "config-data") pod "fcb3e56a-fd7b-4d8e-b42d-2d55de745667" (UID: "fcb3e56a-fd7b-4d8e-b42d-2d55de745667"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500424 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500451 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500461 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50a1fd31-a115-4bb6-b65e-70adc5dd77ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500469 4867 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf84c603-ce1e-42be-a595-b27ddc77880c-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500732 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500789 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts podName:66d1496c-b4da-404b-9f9c-da2c405070da nodeName:}" failed. No retries permitted until 2025-12-12 07:12:40.500774293 +0000 UTC m=+1448.072155562 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts") pod "placementc9b0-account-delete-5g27c" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da") : configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500823 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500854 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500890 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts podName:b500049d-056a-40e5-bce8-b6b9f3196d64 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:40.500866125 +0000 UTC m=+1448.072247454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts") pod "glancea6d9-account-delete-zqpf9" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64") : configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500909 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts podName:cd36e355-181a-4042-a049-aeda5ec0770f nodeName:}" failed. No retries permitted until 2025-12-12 07:12:40.500900836 +0000 UTC m=+1448.072282205 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts") pod "barbican74a1-account-delete-m9cms" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f") : configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500916 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.500941 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts podName:94f74241-b4d4-41fb-b95f-57534e337198 nodeName:}" failed. No retries permitted until 2025-12-12 07:12:40.500930077 +0000 UTC m=+1448.072311426 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts") pod "neutron29f2-account-delete-77p2l" (UID: "94f74241-b4d4-41fb-b95f-57534e337198") : configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.501060 4867 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: E1212 07:12:36.501096 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts podName:b1b2d98f-1a49-4d66-89b3-ca246c9f227c nodeName:}" failed. No retries permitted until 2025-12-12 07:12:40.501086471 +0000 UTC m=+1448.072467810 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts") pod "cinderb8be-account-delete-kvnwc" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c") : configmap "openstack-scripts" not found Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.500477 4867 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501154 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501168 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501178 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501187 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb3e56a-fd7b-4d8e-b42d-2d55de745667-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501208 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501216 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501237 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/746ca9c4-7621-489c-86f2-628a9630039c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501245 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56091088-518c-4636-bb5e-4bdcc9f0397b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501254 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501263 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/163cec12-d32c-4be9-aee0-55703387f2db-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.501271 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.503064 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "437a0623-5c6b-4a29-8da6-f23381fce0ec" (UID: "437a0623-5c6b-4a29-8da6-f23381fce0ec"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.513984 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.521885 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "765da323-2d60-4494-97bc-f699d9bf8661" (UID: "765da323-2d60-4494-97bc-f699d9bf8661"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.527443 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7794fa05-fbb0-4c40-9711-60f299e5ab5b" (UID: "7794fa05-fbb0-4c40-9711-60f299e5ab5b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.533542 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "00269882-fe60-4b26-8d5c-1b34b014b191" (UID: "00269882-fe60-4b26-8d5c-1b34b014b191"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.534990 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "9a11a089-0048-4c70-94e5-9ae18eadfed5" (UID: "9a11a089-0048-4c70-94e5-9ae18eadfed5"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.602945 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.602978 4867 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00269882-fe60-4b26-8d5c-1b34b014b191-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.602989 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a11a089-0048-4c70-94e5-9ae18eadfed5-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.602998 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7794fa05-fbb0-4c40-9711-60f299e5ab5b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.603007 4867 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/437a0623-5c6b-4a29-8da6-f23381fce0ec-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.603018 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/765da323-2d60-4494-97bc-f699d9bf8661-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.777159 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-pdtz8"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.780993 4867 scope.go:117] "RemoveContainer" containerID="9d9541a84d66b2b3172e3884cea7f33b9e6e2538ec678cd2ee81799975bc5831" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.788189 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-pdtz8"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.799921 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c9b0-account-create-update-kmdk9"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.802695 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.857151 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.859140 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e8dc60f-09fa-4c84-b22d-1014fc6d39b4" path="/var/lib/kubelet/pods/2e8dc60f-09fa-4c84-b22d-1014fc6d39b4/volumes" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.859836 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c8e566d-e439-4e4b-bb80-759e9c4d7f89" path="/var/lib/kubelet/pods/3c8e566d-e439-4e4b-bb80-759e9c4d7f89/volumes" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.860455 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" path="/var/lib/kubelet/pods/aa35f883-b75a-4e67-aeed-f36d00a074cf/volumes" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.861967 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" path="/var/lib/kubelet/pods/c7dc7ba4-cc77-4569-ab0f-4bff75ab6472/volumes" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.878529 4867 scope.go:117] "RemoveContainer" containerID="327346d748281830108225a6a9e113b8722507f57f85ac4bbb3af491d43078a5" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913376 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913426 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913459 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913538 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5n8s\" (UniqueName: \"kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913571 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913608 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913650 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913676 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brbn5\" (UniqueName: \"kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913734 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913785 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs\") pod \"b51b1751-7b06-4118-9eff-961ef320bf22\" (UID: \"b51b1751-7b06-4118-9eff-961ef320bf22\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913812 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.913852 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts\") pod \"1ce609f7-1106-4d01-9a6b-d2099112ce33\" (UID: \"1ce609f7-1106-4d01-9a6b-d2099112ce33\") " Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.918677 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.919119 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.920009 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs" (OuterVolumeSpecName: "logs") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.922392 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.922508 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.922847 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.922996 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5" (OuterVolumeSpecName: "kube-api-access-brbn5") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "kube-api-access-brbn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.927476 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s" (OuterVolumeSpecName: "kube-api-access-l5n8s") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "kube-api-access-l5n8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.933387 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c9b0-account-create-update-kmdk9"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.933436 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.956023 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:36 crc kubenswrapper[4867]: I1212 07:12:36.987098 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015158 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015208 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015229 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015276 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015309 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015381 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rh9\" (UniqueName: \"kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015423 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015454 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.015951 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-x8h7k"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016327 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016352 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016939 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brbn5\" (UniqueName: \"kubernetes.io/projected/1ce609f7-1106-4d01-9a6b-d2099112ce33-kube-api-access-brbn5\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016955 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016968 4867 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016979 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce609f7-1106-4d01-9a6b-d2099112ce33-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.016990 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ce609f7-1106-4d01-9a6b-d2099112ce33-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.017002 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5n8s\" (UniqueName: \"kubernetes.io/projected/b51b1751-7b06-4118-9eff-961ef320bf22-kube-api-access-l5n8s\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.017014 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b51b1751-7b06-4118-9eff-961ef320bf22-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.031351 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-x8h7k"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.031487 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.031634 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.040349 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-74a1-account-create-update-mdf7g"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.041429 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.050164 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9" (OuterVolumeSpecName: "kube-api-access-86rh9") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "kube-api-access-86rh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.056215 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-74a1-account-create-update-mdf7g"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.056797 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.066369 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data" (OuterVolumeSpecName: "config-data") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.075145 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts" (OuterVolumeSpecName: "scripts") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.097170 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.107991 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "1ce609f7-1106-4d01-9a6b-d2099112ce33" (UID: "1ce609f7-1106-4d01-9a6b-d2099112ce33"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.115554 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.117973 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data" (OuterVolumeSpecName: "config-data") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119157 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") pod \"38431807-5fbd-4f45-8eca-b42af2c0d221\" (UID: \"38431807-5fbd-4f45-8eca-b42af2c0d221\") " Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.119310 4867 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/38431807-5fbd-4f45-8eca-b42af2c0d221/volumes/kubernetes.io~secret/config-data Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119330 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data" (OuterVolumeSpecName: "config-data") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119924 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119952 4867 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119971 4867 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ce609f7-1106-4d01-9a6b-d2099112ce33-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.119985 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120039 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120069 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120084 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120203 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120278 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.120304 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rh9\" (UniqueName: \"kubernetes.io/projected/38431807-5fbd-4f45-8eca-b42af2c0d221-kube-api-access-86rh9\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.128985 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b51b1751-7b06-4118-9eff-961ef320bf22" (UID: "b51b1751-7b06-4118-9eff-961ef320bf22"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.129357 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.140029 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.140379 4867 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a9d745d_1642_48f4_b1ee_e491f578757e.slice/crio-70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082/memory.min": read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a9d745d_1642_48f4_b1ee_e491f578757e.slice/crio-70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082/memory.min: no such device Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.140943 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice/crio-conmon-00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice/crio-conmon-00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.140980 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d1496c_b4da_404b_9f9c_da2c405070da.slice/crio-conmon-db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d1496c_b4da_404b_9f9c_da2c405070da.slice/crio-conmon-db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.140997 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd36e355_181a_4042_a049_aeda5ec0770f.slice/crio-conmon-1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd36e355_181a_4042_a049_aeda5ec0770f.slice/crio-conmon-1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141019 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94f74241_b4d4_41fb_b95f_57534e337198.slice/crio-conmon-092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94f74241_b4d4_41fb_b95f_57534e337198.slice/crio-conmon-092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141034 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice/crio-conmon-ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice/crio-conmon-ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141063 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d1496c_b4da_404b_9f9c_da2c405070da.slice/crio-db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66d1496c_b4da_404b_9f9c_da2c405070da.slice/crio-db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141117 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice/crio-00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice/crio-00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141142 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94f74241_b4d4_41fb_b95f_57534e337198.slice/crio-092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94f74241_b4d4_41fb_b95f_57534e337198.slice/crio-092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141181 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd36e355_181a_4042_a049_aeda5ec0770f.slice/crio-1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd36e355_181a_4042_a049_aeda5ec0770f.slice/crio-1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: W1212 07:12:37.141203 4867 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice/crio-ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice/crio-ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f.scope: no such file or directory Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.143486 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "38431807-5fbd-4f45-8eca-b42af2c0d221" (UID: "38431807-5fbd-4f45-8eca-b42af2c0d221"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.143645 4867 generic.go:334] "Generic (PLEG): container finished" podID="38431807-5fbd-4f45-8eca-b42af2c0d221" containerID="a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4" exitCode=0 Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.143759 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74796bf6db-754rv" event={"ID":"38431807-5fbd-4f45-8eca-b42af2c0d221","Type":"ContainerDied","Data":"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.143951 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74796bf6db-754rv" event={"ID":"38431807-5fbd-4f45-8eca-b42af2c0d221","Type":"ContainerDied","Data":"6165b97c8c408c0ad06de5b242132547e47ddecd324d40e2800e468f4b8d4f8c"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.143778 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74796bf6db-754rv" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.150923 4867 generic.go:334] "Generic (PLEG): container finished" podID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerID="23284a0c2e794f7efbc84dbb67df86a5355b8c4e5689bfc88bd6bad7d0096872" exitCode=0 Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.151098 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerDied","Data":"23284a0c2e794f7efbc84dbb67df86a5355b8c4e5689bfc88bd6bad7d0096872"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.157716 4867 generic.go:334] "Generic (PLEG): container finished" podID="71501691-562e-4385-a19e-fe9f39cfde61" containerID="0482c34d8c8358e81c2f8bc9ca4911f7c07f4fe5e94574f9a0a5812a8529427f" exitCode=0 Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.157849 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerDied","Data":"0482c34d8c8358e81c2f8bc9ca4911f7c07f4fe5e94574f9a0a5812a8529427f"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.165419 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1ce609f7-1106-4d01-9a6b-d2099112ce33","Type":"ContainerDied","Data":"66da9a2b24b25a23ddff48c1b4c394ad7a15395a4892c18dbf7f7609a63457e3"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.166064 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.161792 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.203:3000/\": dial tcp 10.217.0.203:3000: connect: connection refused" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.193023 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.193083 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b51b1751-7b06-4118-9eff-961ef320bf22","Type":"ContainerDied","Data":"4b978d6c6fa7d736b2906d3a81e5031001373012eaf81a8876153867f05c5c03"} Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.200539 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-znq7k"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.210445 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-znq7k"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.221544 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.221596 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51b1751-7b06-4118-9eff-961ef320bf22-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.221610 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.221624 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38431807-5fbd-4f45-8eca-b42af2c0d221-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.231089 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a6d9-account-create-update-s9hgd"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.238852 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a6d9-account-create-update-s9hgd"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.245179 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.303483 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.303631 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.305747 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.313861 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.317137 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.317316 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.317347 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.344673 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.363777 4867 scope.go:117] "RemoveContainer" containerID="5b52059c4a42740f577bd5d551a5bb09a1e4dd0ae4d8ffaf36f4ba04f2676760" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.369900 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qtct9"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.379815 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:37 crc kubenswrapper[4867]: E1212 07:12:37.379892 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.398358 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qtct9"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.500104 4867 scope.go:117] "RemoveContainer" containerID="82e68b2b9f69ed30e1eee05edc30a399c219b479de169d6d7373d0b615332b11" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.549722 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.557434 4867 scope.go:117] "RemoveContainer" containerID="8cd180ed7cb6b465bdfd0105a35071f5be7a9810e30b308406669748fe930aab" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.601562 4867 scope.go:117] "RemoveContainer" containerID="585c499075ad45a4e241eb7bb0c08e8672b5782b99a85d27b9535747697acf1b" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.613394 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.625461 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b8be-account-create-update-29zkz"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.632638 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.638731 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b8be-account-create-update-29zkz"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.649331 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.657824 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.679262 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.689952 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6f69644fbd-dsggh"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.699844 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.709219 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.717851 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-dz67c"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.727183 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-dz67c"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.742463 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.757884 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8njtj"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.765392 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.773086 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-29f2-account-create-update-qfrmt"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.777281 4867 scope.go:117] "RemoveContainer" containerID="43b9e85411b1cb9f539e0087be32113dcb6485d2c8c7185f67442c9bb1271e17" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.780506 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.788457 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.797747 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-29f2-account-create-update-qfrmt"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.805949 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.813474 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.821254 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.829346 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.836535 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.847606 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.849846 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.855372 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-67c5f5d445-rd5dx"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.861219 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.866103 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.875853 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.881727 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-68ccfcbfb4-q47vm"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.888329 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.888540 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.902359 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.904420 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.908030 4867 scope.go:117] "RemoveContainer" containerID="aa78d51c8939592762c3e6088817a1ab99bc94c48f70defbb2782f5fb28d1164" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.918050 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-74796bf6db-754rv"] Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.918280 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.948301 4867 scope.go:117] "RemoveContainer" containerID="e2a9d914876172e6214a4c0a877ab3a669074c14c1443e7e806e50e7350ff86a" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.968523 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts\") pod \"66d1496c-b4da-404b-9f9c-da2c405070da\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.968586 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrlk6\" (UniqueName: \"kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6\") pod \"cd36e355-181a-4042-a049-aeda5ec0770f\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.968678 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts\") pod \"cd36e355-181a-4042-a049-aeda5ec0770f\" (UID: \"cd36e355-181a-4042-a049-aeda5ec0770f\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.968762 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb2k7\" (UniqueName: \"kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7\") pod \"66d1496c-b4da-404b-9f9c-da2c405070da\" (UID: \"66d1496c-b4da-404b-9f9c-da2c405070da\") " Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.970018 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "66d1496c-b4da-404b-9f9c-da2c405070da" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.970441 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cd36e355-181a-4042-a049-aeda5ec0770f" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.977199 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6" (OuterVolumeSpecName: "kube-api-access-nrlk6") pod "cd36e355-181a-4042-a049-aeda5ec0770f" (UID: "cd36e355-181a-4042-a049-aeda5ec0770f"). InnerVolumeSpecName "kube-api-access-nrlk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:37 crc kubenswrapper[4867]: I1212 07:12:37.980445 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7" (OuterVolumeSpecName: "kube-api-access-sb2k7") pod "66d1496c-b4da-404b-9f9c-da2c405070da" (UID: "66d1496c-b4da-404b-9f9c-da2c405070da"). InnerVolumeSpecName "kube-api-access-sb2k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.071365 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66d1496c-b4da-404b-9f9c-da2c405070da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.071661 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrlk6\" (UniqueName: \"kubernetes.io/projected/cd36e355-181a-4042-a049-aeda5ec0770f-kube-api-access-nrlk6\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.071676 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd36e355-181a-4042-a049-aeda5ec0770f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.071689 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb2k7\" (UniqueName: \"kubernetes.io/projected/66d1496c-b4da-404b-9f9c-da2c405070da-kube-api-access-sb2k7\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.129815 4867 scope.go:117] "RemoveContainer" containerID="ee6e5dae4c9a11595f4e81ee6c481fdc2279b8c9a6898d55722fd9690e003f40" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.160819 4867 scope.go:117] "RemoveContainer" containerID="7af7ea533c4fbeeee8dc600c2255f6879f5ae8111689a0c44bb13bbf800546fe" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.167459 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.167479 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.174897 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.200169 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.222255 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.223031 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican74a1-account-delete-m9cms" event={"ID":"cd36e355-181a-4042-a049-aeda5ec0770f","Type":"ContainerDied","Data":"8413c1fe30ad84dcb710d7a144e960c9c39920973b26eeabf39fdc387b07154c"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.223111 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican74a1-account-delete-m9cms" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.240724 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" event={"ID":"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8","Type":"ContainerDied","Data":"5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.240803 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b4475fc66-dqrh7" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.245467 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancea6d9-account-delete-zqpf9" event={"ID":"b500049d-056a-40e5-bce8-b6b9f3196d64","Type":"ContainerDied","Data":"9dc4fd28ec0f317f12400cdbe7542efe8ef98a5fdf7d59d1a6cd1b0e4a02c823"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.245624 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancea6d9-account-delete-zqpf9" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.248080 4867 generic.go:334] "Generic (PLEG): container finished" podID="00532992-54f5-4a24-b0e4-eb5512e08753" containerID="02be4e6b1d82467e195a35fda03f0c2a9e4fbba01b3f18526ddc7beb6bb54785" exitCode=0 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.248157 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerDied","Data":"02be4e6b1d82467e195a35fda03f0c2a9e4fbba01b3f18526ddc7beb6bb54785"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.263485 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinderb8be-account-delete-kvnwc" event={"ID":"b1b2d98f-1a49-4d66-89b3-ca246c9f227c","Type":"ContainerDied","Data":"22faa0c03d72537d3b22c09966e75ec38e80847b6e5abd135c1b38a198d21c52"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.263747 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinderb8be-account-delete-kvnwc" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.267134 4867 generic.go:334] "Generic (PLEG): container finished" podID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerID="234415f3af38fbe83a304c1655d565aaee048e266eab8c74960cba70dcab3267" exitCode=0 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.267206 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerDied","Data":"234415f3af38fbe83a304c1655d565aaee048e266eab8c74960cba70dcab3267"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.273603 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementc9b0-account-delete-5g27c" event={"ID":"66d1496c-b4da-404b-9f9c-da2c405070da","Type":"ContainerDied","Data":"020f4a25f965ce85a1f6f8b78cc434900900698068d74a3b8073045729e7fd68"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.273765 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementc9b0-account-delete-5g27c" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277185 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle\") pod \"71501691-562e-4385-a19e-fe9f39cfde61\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277362 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277426 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs\") pod \"71501691-562e-4385-a19e-fe9f39cfde61\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277448 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbm8x\" (UniqueName: \"kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277485 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277513 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277530 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle\") pod \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277553 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data\") pod \"71501691-562e-4385-a19e-fe9f39cfde61\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277575 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w29jc\" (UniqueName: \"kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc\") pod \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277599 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom\") pod \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277649 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs\") pod \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277673 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whk44\" (UniqueName: \"kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44\") pod \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277695 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom\") pod \"71501691-562e-4385-a19e-fe9f39cfde61\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277720 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5tfh\" (UniqueName: \"kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh\") pod \"71501691-562e-4385-a19e-fe9f39cfde61\" (UID: \"71501691-562e-4385-a19e-fe9f39cfde61\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277739 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277760 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277819 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts\") pod \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\" (UID: \"b1b2d98f-1a49-4d66-89b3-ca246c9f227c\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277849 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqfsk\" (UniqueName: \"kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk\") pod \"b500049d-056a-40e5-bce8-b6b9f3196d64\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277892 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data\") pod \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\" (UID: \"06e70b8d-199d-4c5b-be9c-7edc6daf4bc8\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277921 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts\") pod \"b500049d-056a-40e5-bce8-b6b9f3196d64\" (UID: \"b500049d-056a-40e5-bce8-b6b9f3196d64\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277948 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.277965 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml\") pod \"4a9d745d-1642-48f4-b1ee-e491f578757e\" (UID: \"4a9d745d-1642-48f4-b1ee-e491f578757e\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.284902 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1b2d98f-1a49-4d66-89b3-ca246c9f227c" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.287591 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.289175 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44" (OuterVolumeSpecName: "kube-api-access-whk44") pod "b1b2d98f-1a49-4d66-89b3-ca246c9f227c" (UID: "b1b2d98f-1a49-4d66-89b3-ca246c9f227c"). InnerVolumeSpecName "kube-api-access-whk44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.290195 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs" (OuterVolumeSpecName: "logs") pod "71501691-562e-4385-a19e-fe9f39cfde61" (UID: "71501691-562e-4385-a19e-fe9f39cfde61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.300510 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "71501691-562e-4385-a19e-fe9f39cfde61" (UID: "71501691-562e-4385-a19e-fe9f39cfde61"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.303116 4867 scope.go:117] "RemoveContainer" containerID="0189962983d6a2e6fe7b73484b68a37d838b9e00becc8036d6e88a5ce8b067ad" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.306295 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh" (OuterVolumeSpecName: "kube-api-access-x5tfh") pod "71501691-562e-4385-a19e-fe9f39cfde61" (UID: "71501691-562e-4385-a19e-fe9f39cfde61"). InnerVolumeSpecName "kube-api-access-x5tfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.307392 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.307659 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b500049d-056a-40e5-bce8-b6b9f3196d64" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.319209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" event={"ID":"71501691-562e-4385-a19e-fe9f39cfde61","Type":"ContainerDied","Data":"9e7b76644371ac1d38436f658179e76b6a5d2acda57f631db86c338a98275947"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.319250 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6d5d4d56b5-6h6kt" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.321730 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs" (OuterVolumeSpecName: "logs") pod "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" (UID: "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.322794 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk" (OuterVolumeSpecName: "kube-api-access-xqfsk") pod "b500049d-056a-40e5-bce8-b6b9f3196d64" (UID: "b500049d-056a-40e5-bce8-b6b9f3196d64"). InnerVolumeSpecName "kube-api-access-xqfsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.328853 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x" (OuterVolumeSpecName: "kube-api-access-gbm8x") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "kube-api-access-gbm8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.330088 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.334619 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" (UID: "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.342508 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerID="363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece" exitCode=0 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.342563 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerDied","Data":"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.342595 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4a9d745d-1642-48f4-b1ee-e491f578757e","Type":"ContainerDied","Data":"70e8353bd844ccb43ed07d8ac386f2c9334db6f68065fff433d3177cb5d81082"} Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.342725 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.346400 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts" (OuterVolumeSpecName: "scripts") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.358124 4867 scope.go:117] "RemoveContainer" containerID="7034fe2f8ab09878f66ad007d5457ffd61d078c6b540144bcfc0cb8f137fc844" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.368373 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.377472 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc" (OuterVolumeSpecName: "kube-api-access-w29jc") pod "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" (UID: "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8"). InnerVolumeSpecName "kube-api-access-w29jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.380786 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2rxk\" (UniqueName: \"kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk\") pod \"94f74241-b4d4-41fb-b95f-57534e337198\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381029 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts\") pod \"94f74241-b4d4-41fb-b95f-57534e337198\" (UID: \"94f74241-b4d4-41fb-b95f-57534e337198\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381416 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71501691-562e-4385-a19e-fe9f39cfde61-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381433 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbm8x\" (UniqueName: \"kubernetes.io/projected/4a9d745d-1642-48f4-b1ee-e491f578757e-kube-api-access-gbm8x\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381443 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381452 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w29jc\" (UniqueName: \"kubernetes.io/projected/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-kube-api-access-w29jc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381460 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381468 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-logs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381477 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whk44\" (UniqueName: \"kubernetes.io/projected/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-kube-api-access-whk44\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381486 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381494 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5tfh\" (UniqueName: \"kubernetes.io/projected/71501691-562e-4385-a19e-fe9f39cfde61-kube-api-access-x5tfh\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381503 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1b2d98f-1a49-4d66-89b3-ca246c9f227c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381511 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqfsk\" (UniqueName: \"kubernetes.io/projected/b500049d-056a-40e5-bce8-b6b9f3196d64-kube-api-access-xqfsk\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381520 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b500049d-056a-40e5-bce8-b6b9f3196d64-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381528 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381536 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4a9d745d-1642-48f4-b1ee-e491f578757e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.381925 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94f74241-b4d4-41fb-b95f-57534e337198" (UID: "94f74241-b4d4-41fb-b95f-57534e337198"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.385371 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican74a1-account-delete-m9cms"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.391772 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.398984 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementc9b0-account-delete-5g27c"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.420513 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71501691-562e-4385-a19e-fe9f39cfde61" (UID: "71501691-562e-4385-a19e-fe9f39cfde61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.420647 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk" (OuterVolumeSpecName: "kube-api-access-t2rxk") pod "94f74241-b4d4-41fb-b95f-57534e337198" (UID: "94f74241-b4d4-41fb-b95f-57534e337198"). InnerVolumeSpecName "kube-api-access-t2rxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.438329 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" (UID: "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.442655 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.450035 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.465087 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data" (OuterVolumeSpecName: "config-data") pod "71501691-562e-4385-a19e-fe9f39cfde61" (UID: "71501691-562e-4385-a19e-fe9f39cfde61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.481359 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data" (OuterVolumeSpecName: "config-data") pod "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" (UID: "06e70b8d-199d-4c5b-be9c-7edc6daf4bc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483482 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483515 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483528 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483540 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94f74241-b4d4-41fb-b95f-57534e337198-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483550 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483562 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483574 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2rxk\" (UniqueName: \"kubernetes.io/projected/94f74241-b4d4-41fb-b95f-57534e337198-kube-api-access-t2rxk\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.483588 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71501691-562e-4385-a19e-fe9f39cfde61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.516235 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.544466 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data" (OuterVolumeSpecName: "config-data") pod "4a9d745d-1642-48f4-b1ee-e491f578757e" (UID: "4a9d745d-1642-48f4-b1ee-e491f578757e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.588391 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.588418 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d745d-1642-48f4-b1ee-e491f578757e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.608294 4867 scope.go:117] "RemoveContainer" containerID="f164d088b657a841b64a09bed96dee589e4f7130244d17533d2c9a7d6fdb6dab" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.642335 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.649665 4867 scope.go:117] "RemoveContainer" containerID="506b0e61cb33a6fbdd013a5cdbf2ecb8a9e984240941a905449f20e6d9bda9a1" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.660970 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.671169 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.677956 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinderb8be-account-delete-kvnwc"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.695050 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.699580 4867 scope.go:117] "RemoveContainer" containerID="a6b387c7b3b91393496b3df6c1f76038666dc95195810690a8e0c383d5e2aa2d" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.704134 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-b4475fc66-dqrh7"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.726772 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.732933 4867 scope.go:117] "RemoveContainer" containerID="7d0afa62d13fd39fc3b3ba841aff538f37e9ec010eeac42f3c84ad166e59f590" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.753172 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.772505 4867 scope.go:117] "RemoveContainer" containerID="dc60c646092d5fd6e48768335070dbf4389a632716a9616ea70be7334134e290" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.779249 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797312 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797395 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797434 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797460 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn89r\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797486 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797506 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jtzq\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797531 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797558 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797594 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797618 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797642 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797672 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797702 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797718 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797732 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797768 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797792 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797832 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797849 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797867 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797887 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.797905 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret\") pod \"00532992-54f5-4a24-b0e4-eb5512e08753\" (UID: \"00532992-54f5-4a24-b0e4-eb5512e08753\") " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.803403 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.809483 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.822628 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.828175 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.828636 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.828898 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.833148 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancea6d9-account-delete-zqpf9"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.834665 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.834913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.835256 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: E1212 07:12:38.843292 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06e70b8d_199d_4c5b_be9c_7edc6daf4bc8.slice/crio-5994ba60d1437017be173cfeffc755acca0104b5a8c34712b87198af86e2fbee\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb500049d_056a_40e5_bce8_b6b9f3196d64.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71501691_562e_4385_a19e_fe9f39cfde61.slice/crio-9e7b76644371ac1d38436f658179e76b6a5d2acda57f631db86c338a98275947\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71501691_562e_4385_a19e_fe9f39cfde61.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice/crio-22faa0c03d72537d3b22c09966e75ec38e80847b6e5abd135c1b38a198d21c52\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1b2d98f_1a49_4d66_89b3_ca246c9f227c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a9d745d_1642_48f4_b1ee_e491f578757e.slice\": RecentStats: unable to find data in memory cache]" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.854341 4867 scope.go:117] "RemoveContainer" containerID="bc30a378f2f47b54d0e25ab70eab05c43f7eb5652e82cd1c2e5b19b297703f21" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.869929 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info" (OuterVolumeSpecName: "pod-info") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.870300 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" path="/var/lib/kubelet/pods/00269882-fe60-4b26-8d5c-1b34b014b191/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.877385 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" path="/var/lib/kubelet/pods/06e70b8d-199d-4c5b-be9c-7edc6daf4bc8/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.878625 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r" (OuterVolumeSpecName: "kube-api-access-cn89r") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "kube-api-access-cn89r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.878750 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.881748 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163cec12-d32c-4be9-aee0-55703387f2db" path="/var/lib/kubelet/pods/163cec12-d32c-4be9-aee0-55703387f2db/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.885011 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" path="/var/lib/kubelet/pods/1ce609f7-1106-4d01-9a6b-d2099112ce33/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.888545 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38431807-5fbd-4f45-8eca-b42af2c0d221" path="/var/lib/kubelet/pods/38431807-5fbd-4f45-8eca-b42af2c0d221/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.889624 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437a0623-5c6b-4a29-8da6-f23381fce0ec" path="/var/lib/kubelet/pods/437a0623-5c6b-4a29-8da6-f23381fce0ec/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.890245 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" path="/var/lib/kubelet/pods/4a9d745d-1642-48f4-b1ee-e491f578757e/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.891462 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq" (OuterVolumeSpecName: "kube-api-access-9jtzq") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "kube-api-access-9jtzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.891472 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.891665 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info" (OuterVolumeSpecName: "pod-info") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.891692 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.892138 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" path="/var/lib/kubelet/pods/50a1fd31-a115-4bb6-b65e-70adc5dd77ed/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.893457 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" path="/var/lib/kubelet/pods/56091088-518c-4636-bb5e-4bdcc9f0397b/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.895530 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5986e88e-5488-4faa-829f-eebfbf81f209" path="/var/lib/kubelet/pods/5986e88e-5488-4faa-829f-eebfbf81f209/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.896263 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" path="/var/lib/kubelet/pods/66d1496c-b4da-404b-9f9c-da2c405070da/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.897729 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="746ca9c4-7621-489c-86f2-628a9630039c" path="/var/lib/kubelet/pods/746ca9c4-7621-489c-86f2-628a9630039c/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.898267 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="765da323-2d60-4494-97bc-f699d9bf8661" path="/var/lib/kubelet/pods/765da323-2d60-4494-97bc-f699d9bf8661/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.901905 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" path="/var/lib/kubelet/pods/7794fa05-fbb0-4c40-9711-60f299e5ab5b/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.902764 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ec53377-bb33-44d4-af94-7b24177ff046" path="/var/lib/kubelet/pods/7ec53377-bb33-44d4-af94-7b24177ff046/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903153 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903178 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903188 4867 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/00532992-54f5-4a24-b0e4-eb5512e08753-pod-info\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903198 4867 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903207 4867 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/65ad43b4-b2fa-4089-b471-66fbae81fbf2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903216 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903226 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903305 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dcfd4ee-cde3-49e8-ad50-af6fd23af517" path="/var/lib/kubelet/pods/8dcfd4ee-cde3-49e8-ad50-af6fd23af517/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903336 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903347 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903358 4867 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/00532992-54f5-4a24-b0e4-eb5512e08753-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903370 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903379 4867 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/65ad43b4-b2fa-4089-b471-66fbae81fbf2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903387 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn89r\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-kube-api-access-cn89r\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903395 4867 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903403 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jtzq\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-kube-api-access-9jtzq\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.903412 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.904266 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" path="/var/lib/kubelet/pods/9a11a089-0048-4c70-94e5-9ae18eadfed5/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.904868 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3" path="/var/lib/kubelet/pods/a3ba780e-0b5f-4e6f-9250-2b1eff1cbec3/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.905400 4867 scope.go:117] "RemoveContainer" containerID="35dfd871e3575c594368d88c9965b5034e1b0024ed10fec18e5ec7de472a25cf" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.905429 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" path="/var/lib/kubelet/pods/b1b2d98f-1a49-4d66-89b3-ca246c9f227c/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.906387 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" path="/var/lib/kubelet/pods/b500049d-056a-40e5-bce8-b6b9f3196d64/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.906933 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" path="/var/lib/kubelet/pods/b51b1751-7b06-4118-9eff-961ef320bf22/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.907504 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9fabde1-97a0-4d14-8613-6bd031ab968d" path="/var/lib/kubelet/pods/b9fabde1-97a0-4d14-8613-6bd031ab968d/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.908475 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" path="/var/lib/kubelet/pods/bf84c603-ce1e-42be-a595-b27ddc77880c/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.909009 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8faa97c-b754-433d-ae7c-d875a0cd286e" path="/var/lib/kubelet/pods/c8faa97c-b754-433d-ae7c-d875a0cd286e/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.909523 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cba96708-8c06-4d68-bd49-1cf930cc5cfb" path="/var/lib/kubelet/pods/cba96708-8c06-4d68-bd49-1cf930cc5cfb/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.910041 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" path="/var/lib/kubelet/pods/cd36e355-181a-4042-a049-aeda5ec0770f/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.915403 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data" (OuterVolumeSpecName: "config-data") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.918412 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data" (OuterVolumeSpecName: "config-data") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.920141 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e4e236-bbc3-42bd-8c73-3386ff4556ba" path="/var/lib/kubelet/pods/d2e4e236-bbc3-42bd-8c73-3386ff4556ba/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.924100 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.924787 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6da4f5f-8f05-47d2-a3d5-79cfd7666d51" path="/var/lib/kubelet/pods/e6da4f5f-8f05-47d2-a3d5-79cfd7666d51/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.925369 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" path="/var/lib/kubelet/pods/fbb7b7c4-4c08-4f29-8e06-18af9f7422ae/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.925919 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" path="/var/lib/kubelet/pods/fcb3e56a-fd7b-4d8e-b42d-2d55de745667/volumes" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.933909 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.941408 4867 scope.go:117] "RemoveContainer" containerID="a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.941676 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.942657 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-6d5d4d56b5-6h6kt"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943384 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943406 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943419 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dlfmg"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943431 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dlfmg"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943440 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943451 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xkgvg"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943462 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xkgvg"] Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.943652 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" containerName="nova-cell0-conductor-conductor" containerID="cri-o://848266a1cfe8cee63f4a2009ad41e8a49241ba1784825fa11ad76c5f036dc5c0" gracePeriod=30 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.944002 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="45fe5995-c9da-4b48-bef1-20cc51f101f2" containerName="memcached" containerID="cri-o://b4b21d46b7b68ccba3a77319862cb744cd50115fe97c79196a5c300c08368180" gracePeriod=30 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.944123 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="366342dd-c366-4639-9f6d-6444f38fed68" containerName="nova-cell1-conductor-conductor" containerID="cri-o://00a292211a141c686e94c55059b96c051c4cc645dd2e169e1b511c3b048cd336" gracePeriod=30 Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.964159 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf" (OuterVolumeSpecName: "server-conf") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.973535 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf" (OuterVolumeSpecName: "server-conf") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.980054 4867 scope.go:117] "RemoveContainer" containerID="a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4" Dec 12 07:12:38 crc kubenswrapper[4867]: E1212 07:12:38.981192 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4\": container with ID starting with a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4 not found: ID does not exist" containerID="a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.981503 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4"} err="failed to get container status \"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4\": rpc error: code = NotFound desc = could not find container \"a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4\": container with ID starting with a585edbb80f53e9c15924c2852546e61367755ce7d8dd0c4bd99e27b7e309ac4 not found: ID does not exist" Dec 12 07:12:38 crc kubenswrapper[4867]: I1212 07:12:38.981543 4867 scope.go:117] "RemoveContainer" containerID="49e671f7afaeff4def9023190941487f05677eb906ef9b83564e24d0035e039e" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.004537 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.004805 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") pod \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\" (UID: \"65ad43b4-b2fa-4089-b471-66fbae81fbf2\") " Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005130 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005144 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005152 4867 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/00532992-54f5-4a24-b0e4-eb5512e08753-server-conf\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005160 4867 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/65ad43b4-b2fa-4089-b471-66fbae81fbf2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005169 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005178 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: W1212 07:12:39.005276 4867 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/65ad43b4-b2fa-4089-b471-66fbae81fbf2/volumes/kubernetes.io~projected/rabbitmq-confd Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.005337 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "65ad43b4-b2fa-4089-b471-66fbae81fbf2" (UID: "65ad43b4-b2fa-4089-b471-66fbae81fbf2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.010526 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "00532992-54f5-4a24-b0e4-eb5512e08753" (UID: "00532992-54f5-4a24-b0e4-eb5512e08753"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.018680 4867 scope.go:117] "RemoveContainer" containerID="e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.040733 4867 scope.go:117] "RemoveContainer" containerID="0f947d9d4fe9c6b4d937ad0d0687b9d1db4c6172ed3a4aac66489775077c3e68" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.060162 4867 scope.go:117] "RemoveContainer" containerID="b6413fee15f4b9393e64a3a1314ffb75d76cc3621c445d54a57ddbb7d9e78327" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.084787 4867 scope.go:117] "RemoveContainer" containerID="95bf09c837b62e7939b11139065bedf6b217dfd70415ec0d948de99abeef31dc" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.102850 4867 scope.go:117] "RemoveContainer" containerID="f7cd3a86953862b945327d07440fe89481f1546ab42431b9a9517408da200482" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.107385 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/65ad43b4-b2fa-4089-b471-66fbae81fbf2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.107422 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/00532992-54f5-4a24-b0e4-eb5512e08753-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.116698 4867 scope.go:117] "RemoveContainer" containerID="d4f327c63f9eac35349fd388c60155290ee649c3a89069eef1b3a6f6505d6d1f" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.127610 4867 scope.go:117] "RemoveContainer" containerID="1519ee5fa22316c492fb1d5a60eb6c4ff39d030202cddb0c73d648c49baa780e" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.182432 4867 scope.go:117] "RemoveContainer" containerID="48c957116fe30fd7ec609dae774ee778a0db6e286342964dbcc710c0a4645098" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.218316 4867 scope.go:117] "RemoveContainer" containerID="23284a0c2e794f7efbc84dbb67df86a5355b8c4e5689bfc88bd6bad7d0096872" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.241953 4867 scope.go:117] "RemoveContainer" containerID="e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.258573 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e\": container with ID starting with e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e not found: ID does not exist" containerID="e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.258627 4867 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e\": rpc error: code = NotFound desc = could not find container \"e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e\": container with ID starting with e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e not found: ID does not exist" containerID="e3ea160c3b6df3cfd203a6e823d22500160dcd082f664e589190f21f5111085e" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.258662 4867 scope.go:117] "RemoveContainer" containerID="ff33b41e33f666b184b75128290757493fa986e1b0ac7b622f277504d200fe71" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.313200 4867 scope.go:117] "RemoveContainer" containerID="14a5c4be4116f44371426de020f2bddb11420dd71b7ecd0f938dff04595d0193" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.328070 4867 scope.go:117] "RemoveContainer" containerID="de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.357666 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"00532992-54f5-4a24-b0e4-eb5512e08753","Type":"ContainerDied","Data":"cfbbc9e4351cbf29e8bafa83fd622c0cc562bf053e29dadb8d8188072ae0f4ec"} Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.357743 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.367612 4867 generic.go:334] "Generic (PLEG): container finished" podID="5e79e315-4744-4806-96fc-284d97fa88e0" containerID="456526542b9c5797b244ddfdcbdceb1b0f47fb01b1e916fbb187836108af239b" exitCode=0 Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.367688 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerDied","Data":"456526542b9c5797b244ddfdcbdceb1b0f47fb01b1e916fbb187836108af239b"} Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.392284 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron29f2-account-delete-77p2l" event={"ID":"94f74241-b4d4-41fb-b95f-57534e337198","Type":"ContainerDied","Data":"c79730947610ada694194883b4630e38707000fca4d36cb10a6f0ad33476e446"} Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.392398 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron29f2-account-delete-77p2l" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.397824 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"65ad43b4-b2fa-4089-b471-66fbae81fbf2","Type":"ContainerDied","Data":"bf7d57e9df490ee931e5c05133417a009be1edebd8eca9d810067ef809aee181"} Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.397918 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.484315 4867 scope.go:117] "RemoveContainer" containerID="ad46167e1ef49c329a682c470a0d7b445a74fc8d245c64bae44787205c2aa02f" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.495771 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.505827 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron29f2-account-delete-77p2l"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.510884 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.515161 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.536716 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.542078 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.575493 4867 scope.go:117] "RemoveContainer" containerID="00ea102610eaba488fc6fe65e69157f1c4b209378b8a6d5591737fb4ab4f18a3" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.598678 4867 scope.go:117] "RemoveContainer" containerID="db151cba30d08497d2a5baa64fb808dc0dff9150f8f8a0226eed187e3eaafd6e" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.658910 4867 scope.go:117] "RemoveContainer" containerID="0482c34d8c8358e81c2f8bc9ca4911f7c07f4fe5e94574f9a0a5812a8529427f" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.694654 4867 scope.go:117] "RemoveContainer" containerID="1499f0119bb5be4db9007f30932c9a39e2288f464937e17ccb75934aea9e9680" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.730208 4867 scope.go:117] "RemoveContainer" containerID="eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.755109 4867 scope.go:117] "RemoveContainer" containerID="6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.790516 4867 scope.go:117] "RemoveContainer" containerID="363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.809285 4867 scope.go:117] "RemoveContainer" containerID="b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.831325 4867 scope.go:117] "RemoveContainer" containerID="eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.831949 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b\": container with ID starting with eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b not found: ID does not exist" containerID="eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.832002 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b"} err="failed to get container status \"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b\": rpc error: code = NotFound desc = could not find container \"eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b\": container with ID starting with eec88b4592676981476b6af7e72ebbee67c0e1321ed4d442821ac4021daf266b not found: ID does not exist" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.832036 4867 scope.go:117] "RemoveContainer" containerID="6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.832498 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474\": container with ID starting with 6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474 not found: ID does not exist" containerID="6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.832523 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474"} err="failed to get container status \"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474\": rpc error: code = NotFound desc = could not find container \"6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474\": container with ID starting with 6310958bc7ce32fbbaa338a9ce5993df4c4c6b050bfea77d7d0e942f01b5c474 not found: ID does not exist" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.832541 4867 scope.go:117] "RemoveContainer" containerID="363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.832961 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece\": container with ID starting with 363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece not found: ID does not exist" containerID="363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.832993 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece"} err="failed to get container status \"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece\": rpc error: code = NotFound desc = could not find container \"363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece\": container with ID starting with 363f1dbdf66d2b43cda9ede3b2306d9a421a74ec208d07a3e349f6722802aece not found: ID does not exist" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.833019 4867 scope.go:117] "RemoveContainer" containerID="b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.833436 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a\": container with ID starting with b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a not found: ID does not exist" containerID="b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.833499 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a"} err="failed to get container status \"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a\": rpc error: code = NotFound desc = could not find container \"b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a\": container with ID starting with b14533a38b80cc6e18b0572cd210b81a3c298ecff8776de79d3cc78a05c2d28a not found: ID does not exist" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.833529 4867 scope.go:117] "RemoveContainer" containerID="02be4e6b1d82467e195a35fda03f0c2a9e4fbba01b3f18526ddc7beb6bb54785" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.877472 4867 scope.go:117] "RemoveContainer" containerID="092464745e777ad4c8d37a6ddf192f4d00cdcc819204cca017814d1103307ec0" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.925678 4867 scope.go:117] "RemoveContainer" containerID="234415f3af38fbe83a304c1655d565aaee048e266eab8c74960cba70dcab3267" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.967173 4867 scope.go:117] "RemoveContainer" containerID="de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020" Dec 12 07:12:39 crc kubenswrapper[4867]: E1212 07:12:39.967705 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020\": container with ID starting with de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020 not found: ID does not exist" containerID="de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020" Dec 12 07:12:39 crc kubenswrapper[4867]: I1212 07:12:39.967735 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020"} err="failed to get container status \"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020\": rpc error: code = NotFound desc = could not find container \"de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020\": container with ID starting with de430dbf94ba305f5a98eaf0b0a40cd4bc0f7e84f198c44cda6187ba5f9f7020 not found: ID does not exist" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.446021 4867 generic.go:334] "Generic (PLEG): container finished" podID="366342dd-c366-4639-9f6d-6444f38fed68" containerID="00a292211a141c686e94c55059b96c051c4cc645dd2e169e1b511c3b048cd336" exitCode=0 Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.446128 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"366342dd-c366-4639-9f6d-6444f38fed68","Type":"ContainerDied","Data":"00a292211a141c686e94c55059b96c051c4cc645dd2e169e1b511c3b048cd336"} Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.453394 4867 generic.go:334] "Generic (PLEG): container finished" podID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" containerID="848266a1cfe8cee63f4a2009ad41e8a49241ba1784825fa11ad76c5f036dc5c0" exitCode=0 Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.453469 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f","Type":"ContainerDied","Data":"848266a1cfe8cee63f4a2009ad41e8a49241ba1784825fa11ad76c5f036dc5c0"} Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.455885 4867 generic.go:334] "Generic (PLEG): container finished" podID="45fe5995-c9da-4b48-bef1-20cc51f101f2" containerID="b4b21d46b7b68ccba3a77319862cb744cd50115fe97c79196a5c300c08368180" exitCode=0 Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.455944 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"45fe5995-c9da-4b48-bef1-20cc51f101f2","Type":"ContainerDied","Data":"b4b21d46b7b68ccba3a77319862cb744cd50115fe97c79196a5c300c08368180"} Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.552997 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: i/o timeout" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.553384 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6f69644fbd-dsggh" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.151:9311/healthcheck\": dial tcp 10.217.0.151:9311: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.743574 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.806797 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.812814 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835140 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgvv9\" (UniqueName: \"kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835187 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835260 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835299 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835318 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835341 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.835420 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts\") pod \"5e79e315-4744-4806-96fc-284d97fa88e0\" (UID: \"5e79e315-4744-4806-96fc-284d97fa88e0\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.836031 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.836459 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.836607 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.838606 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.843038 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9" (OuterVolumeSpecName: "kube-api-access-sgvv9") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "kube-api-access-sgvv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.853571 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.861075 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" path="/var/lib/kubelet/pods/00532992-54f5-4a24-b0e4-eb5512e08753/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.861742 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b73a04-c864-44ed-b88d-9b6a5073861a" path="/var/lib/kubelet/pods/44b73a04-c864-44ed-b88d-9b6a5073861a/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.862478 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" path="/var/lib/kubelet/pods/65ad43b4-b2fa-4089-b471-66fbae81fbf2/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.863794 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="714a3a25-e3f2-4847-9200-0c633bc0254f" path="/var/lib/kubelet/pods/714a3a25-e3f2-4847-9200-0c633bc0254f/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.864332 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71501691-562e-4385-a19e-fe9f39cfde61" path="/var/lib/kubelet/pods/71501691-562e-4385-a19e-fe9f39cfde61/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.865343 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94f74241-b4d4-41fb-b95f-57534e337198" path="/var/lib/kubelet/pods/94f74241-b4d4-41fb-b95f-57534e337198/volumes" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.880550 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.896435 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "5e79e315-4744-4806-96fc-284d97fa88e0" (UID: "5e79e315-4744-4806-96fc-284d97fa88e0"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.936590 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data\") pod \"366342dd-c366-4639-9f6d-6444f38fed68\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.936668 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data\") pod \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937141 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rjs6\" (UniqueName: \"kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6\") pod \"366342dd-c366-4639-9f6d-6444f38fed68\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937244 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle\") pod \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937301 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9mn9\" (UniqueName: \"kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9\") pod \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\" (UID: \"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937401 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle\") pod \"366342dd-c366-4639-9f6d-6444f38fed68\" (UID: \"366342dd-c366-4639-9f6d-6444f38fed68\") " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937859 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937874 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937885 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgvv9\" (UniqueName: \"kubernetes.io/projected/5e79e315-4744-4806-96fc-284d97fa88e0-kube-api-access-sgvv9\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937897 4867 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5e79e315-4744-4806-96fc-284d97fa88e0-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937916 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937948 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5e79e315-4744-4806-96fc-284d97fa88e0-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937958 4867 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.937967 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79e315-4744-4806-96fc-284d97fa88e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.941031 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6" (OuterVolumeSpecName: "kube-api-access-8rjs6") pod "366342dd-c366-4639-9f6d-6444f38fed68" (UID: "366342dd-c366-4639-9f6d-6444f38fed68"). InnerVolumeSpecName "kube-api-access-8rjs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.941825 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9" (OuterVolumeSpecName: "kube-api-access-k9mn9") pod "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" (UID: "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f"). InnerVolumeSpecName "kube-api-access-k9mn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.957654 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.966613 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data" (OuterVolumeSpecName: "config-data") pod "366342dd-c366-4639-9f6d-6444f38fed68" (UID: "366342dd-c366-4639-9f6d-6444f38fed68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.967577 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "366342dd-c366-4639-9f6d-6444f38fed68" (UID: "366342dd-c366-4639-9f6d-6444f38fed68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.968371 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" (UID: "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.970965 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 07:12:40 crc kubenswrapper[4867]: I1212 07:12:40.976558 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data" (OuterVolumeSpecName: "config-data") pod "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" (UID: "4f784fec-6124-4aa4-9de3-5d4fd4d58f3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.039329 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg8ql\" (UniqueName: \"kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql\") pod \"45fe5995-c9da-4b48-bef1-20cc51f101f2\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.039417 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data\") pod \"45fe5995-c9da-4b48-bef1-20cc51f101f2\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.039471 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config\") pod \"45fe5995-c9da-4b48-bef1-20cc51f101f2\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.039501 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle\") pod \"45fe5995-c9da-4b48-bef1-20cc51f101f2\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.039528 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs\") pod \"45fe5995-c9da-4b48-bef1-20cc51f101f2\" (UID: \"45fe5995-c9da-4b48-bef1-20cc51f101f2\") " Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040081 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040098 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rjs6\" (UniqueName: \"kubernetes.io/projected/366342dd-c366-4639-9f6d-6444f38fed68-kube-api-access-8rjs6\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040112 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040127 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040139 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9mn9\" (UniqueName: \"kubernetes.io/projected/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f-kube-api-access-k9mn9\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040150 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.040161 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366342dd-c366-4639-9f6d-6444f38fed68-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.041085 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data" (OuterVolumeSpecName: "config-data") pod "45fe5995-c9da-4b48-bef1-20cc51f101f2" (UID: "45fe5995-c9da-4b48-bef1-20cc51f101f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.041645 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "45fe5995-c9da-4b48-bef1-20cc51f101f2" (UID: "45fe5995-c9da-4b48-bef1-20cc51f101f2"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.045380 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql" (OuterVolumeSpecName: "kube-api-access-xg8ql") pod "45fe5995-c9da-4b48-bef1-20cc51f101f2" (UID: "45fe5995-c9da-4b48-bef1-20cc51f101f2"). InnerVolumeSpecName "kube-api-access-xg8ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.062598 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45fe5995-c9da-4b48-bef1-20cc51f101f2" (UID: "45fe5995-c9da-4b48-bef1-20cc51f101f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.107696 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "45fe5995-c9da-4b48-bef1-20cc51f101f2" (UID: "45fe5995-c9da-4b48-bef1-20cc51f101f2"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.141809 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.141847 4867 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/45fe5995-c9da-4b48-bef1-20cc51f101f2-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.141856 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.141867 4867 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/45fe5995-c9da-4b48-bef1-20cc51f101f2-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.141877 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg8ql\" (UniqueName: \"kubernetes.io/projected/45fe5995-c9da-4b48-bef1-20cc51f101f2-kube-api-access-xg8ql\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.471706 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5e79e315-4744-4806-96fc-284d97fa88e0","Type":"ContainerDied","Data":"7a3352d8bd3e4131024a0604a55fee9457316c1ddd8d1eeff2808f0dd1f89a30"} Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.471733 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.471766 4867 scope.go:117] "RemoveContainer" containerID="456526542b9c5797b244ddfdcbdceb1b0f47fb01b1e916fbb187836108af239b" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.474158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"45fe5995-c9da-4b48-bef1-20cc51f101f2","Type":"ContainerDied","Data":"3105b7f04980db6866b53c011e49ea9c426dd90bf4770bf93d92474332028e8b"} Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.474205 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.476315 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"366342dd-c366-4639-9f6d-6444f38fed68","Type":"ContainerDied","Data":"abf715b11364c9206fbaf205c1a4c22aae12ace87b8d0a6d051596c0030a34c1"} Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.476338 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.478397 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4f784fec-6124-4aa4-9de3-5d4fd4d58f3f","Type":"ContainerDied","Data":"78efefccff02d40e18cef31a2da4391686812283e0342c0595524e040fbdeb99"} Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.478505 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.506305 4867 scope.go:117] "RemoveContainer" containerID="e8768b0eadae644cd43b37f9b16a6da9182b19bd8b06c8288046607848e24de1" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.517278 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.531486 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.536278 4867 scope.go:117] "RemoveContainer" containerID="b4b21d46b7b68ccba3a77319862cb744cd50115fe97c79196a5c300c08368180" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.560752 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.574198 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.586383 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.591003 4867 scope.go:117] "RemoveContainer" containerID="00a292211a141c686e94c55059b96c051c4cc645dd2e169e1b511c3b048cd336" Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.594401 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.602270 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.607308 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 12 07:12:41 crc kubenswrapper[4867]: I1212 07:12:41.611324 4867 scope.go:117] "RemoveContainer" containerID="848266a1cfe8cee63f4a2009ad41e8a49241ba1784825fa11ad76c5f036dc5c0" Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.292092 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.292454 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.292589 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.292891 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.292921 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.293780 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.294877 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:42 crc kubenswrapper[4867]: E1212 07:12:42.294914 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:42 crc kubenswrapper[4867]: I1212 07:12:42.853349 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366342dd-c366-4639-9f6d-6444f38fed68" path="/var/lib/kubelet/pods/366342dd-c366-4639-9f6d-6444f38fed68/volumes" Dec 12 07:12:42 crc kubenswrapper[4867]: I1212 07:12:42.854404 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45fe5995-c9da-4b48-bef1-20cc51f101f2" path="/var/lib/kubelet/pods/45fe5995-c9da-4b48-bef1-20cc51f101f2/volumes" Dec 12 07:12:42 crc kubenswrapper[4867]: I1212 07:12:42.854908 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" path="/var/lib/kubelet/pods/4f784fec-6124-4aa4-9de3-5d4fd4d58f3f/volumes" Dec 12 07:12:42 crc kubenswrapper[4867]: I1212 07:12:42.856881 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" path="/var/lib/kubelet/pods/5e79e315-4744-4806-96fc-284d97fa88e0/volumes" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.225706 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303095 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303180 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303210 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2mb7\" (UniqueName: \"kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303277 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303371 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303390 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.303462 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config\") pod \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\" (UID: \"e267fbd1-d3c8-4379-bb05-e9994fbb13f2\") " Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.309239 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7" (OuterVolumeSpecName: "kube-api-access-z2mb7") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "kube-api-access-z2mb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.318096 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.336930 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.343753 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.344517 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.377111 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config" (OuterVolumeSpecName: "config") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.384984 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e267fbd1-d3c8-4379-bb05-e9994fbb13f2" (UID: "e267fbd1-d3c8-4379-bb05-e9994fbb13f2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405287 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405329 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405341 4867 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405350 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2mb7\" (UniqueName: \"kubernetes.io/projected/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-kube-api-access-z2mb7\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405358 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405367 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.405375 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e267fbd1-d3c8-4379-bb05-e9994fbb13f2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.516388 4867 generic.go:334] "Generic (PLEG): container finished" podID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerID="83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b" exitCode=0 Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.516453 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6796c56757-d2snl" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.516454 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerDied","Data":"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b"} Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.516547 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6796c56757-d2snl" event={"ID":"e267fbd1-d3c8-4379-bb05-e9994fbb13f2","Type":"ContainerDied","Data":"6977f05a7db0225dce26c456a746ae79bf41e55e65070e03c867704eba3d3f00"} Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.516612 4867 scope.go:117] "RemoveContainer" containerID="7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.548646 4867 scope.go:117] "RemoveContainer" containerID="83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.550777 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.572089 4867 scope.go:117] "RemoveContainer" containerID="7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478" Dec 12 07:12:45 crc kubenswrapper[4867]: E1212 07:12:45.573015 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478\": container with ID starting with 7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478 not found: ID does not exist" containerID="7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.573715 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478"} err="failed to get container status \"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478\": rpc error: code = NotFound desc = could not find container \"7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478\": container with ID starting with 7ab4fc16db0ce19a83e9624ed315e6188bcaf55dbe3c71edfe84d6e327597478 not found: ID does not exist" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.573757 4867 scope.go:117] "RemoveContainer" containerID="83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b" Dec 12 07:12:45 crc kubenswrapper[4867]: E1212 07:12:45.574122 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b\": container with ID starting with 83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b not found: ID does not exist" containerID="83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.574146 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b"} err="failed to get container status \"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b\": rpc error: code = NotFound desc = could not find container \"83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b\": container with ID starting with 83dab4a03f9ed493f8fa31ad8540aa2b8eee55e3590cbc5919c7634616d0df6b not found: ID does not exist" Dec 12 07:12:45 crc kubenswrapper[4867]: I1212 07:12:45.575101 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6796c56757-d2snl"] Dec 12 07:12:46 crc kubenswrapper[4867]: I1212 07:12:46.847624 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" path="/var/lib/kubelet/pods/e267fbd1-d3c8-4379-bb05-e9994fbb13f2/volumes" Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.291069 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.291417 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.292151 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.292194 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.293165 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.294580 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.296134 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:47 crc kubenswrapper[4867]: E1212 07:12:47.296168 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.291969 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.293125 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.293503 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.293549 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.293895 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.295507 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.297388 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:52 crc kubenswrapper[4867]: E1212 07:12:52.297448 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.292295 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.292709 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.293107 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.294795 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.294838 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.295068 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.296217 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 12 07:12:57 crc kubenswrapper[4867]: E1212 07:12:57.296272 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-9k5zx" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.531909 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591062 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache\") pod \"786b1010-d0a8-4c94-a42c-99226ab568be\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591420 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") pod \"786b1010-d0a8-4c94-a42c-99226ab568be\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591478 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbqfq\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq\") pod \"786b1010-d0a8-4c94-a42c-99226ab568be\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591502 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"786b1010-d0a8-4c94-a42c-99226ab568be\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591580 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock\") pod \"786b1010-d0a8-4c94-a42c-99226ab568be\" (UID: \"786b1010-d0a8-4c94-a42c-99226ab568be\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.591937 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache" (OuterVolumeSpecName: "cache") pod "786b1010-d0a8-4c94-a42c-99226ab568be" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.592200 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock" (OuterVolumeSpecName: "lock") pod "786b1010-d0a8-4c94-a42c-99226ab568be" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.597801 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "786b1010-d0a8-4c94-a42c-99226ab568be" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.597888 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq" (OuterVolumeSpecName: "kube-api-access-nbqfq") pod "786b1010-d0a8-4c94-a42c-99226ab568be" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be"). InnerVolumeSpecName "kube-api-access-nbqfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.598945 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "swift") pod "786b1010-d0a8-4c94-a42c-99226ab568be" (UID: "786b1010-d0a8-4c94-a42c-99226ab568be"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.651973 4867 generic.go:334] "Generic (PLEG): container finished" podID="786b1010-d0a8-4c94-a42c-99226ab568be" containerID="09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb" exitCode=137 Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.652034 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb"} Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.652103 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"786b1010-d0a8-4c94-a42c-99226ab568be","Type":"ContainerDied","Data":"b09884bcfe897aaae2d5127126df71dbd8b6496dc990b7b03cfd3c2314f7431a"} Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.652125 4867 scope.go:117] "RemoveContainer" containerID="09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.652133 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.656659 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9k5zx_887e1080-2cf6-4c02-a750-b9498fb2f2c0/ovs-vswitchd/0.log" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.657301 4867 generic.go:334] "Generic (PLEG): container finished" podID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" exitCode=137 Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.657350 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerDied","Data":"fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11"} Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.681414 4867 scope.go:117] "RemoveContainer" containerID="3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.687872 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.693388 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.694208 4867 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-cache\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.694251 4867 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.694267 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbqfq\" (UniqueName: \"kubernetes.io/projected/786b1010-d0a8-4c94-a42c-99226ab568be-kube-api-access-nbqfq\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.694308 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.694323 4867 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/786b1010-d0a8-4c94-a42c-99226ab568be-lock\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.703509 4867 scope.go:117] "RemoveContainer" containerID="517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.716750 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.721533 4867 scope.go:117] "RemoveContainer" containerID="b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.748492 4867 scope.go:117] "RemoveContainer" containerID="b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.767129 4867 scope.go:117] "RemoveContainer" containerID="6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.789178 4867 scope.go:117] "RemoveContainer" containerID="1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.789238 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9k5zx_887e1080-2cf6-4c02-a750-b9498fb2f2c0/ovs-vswitchd/0.log" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.790139 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.795840 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.805550 4867 scope.go:117] "RemoveContainer" containerID="572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.829172 4867 scope.go:117] "RemoveContainer" containerID="bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.845419 4867 scope.go:117] "RemoveContainer" containerID="6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.848460 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" path="/var/lib/kubelet/pods/786b1010-d0a8-4c94-a42c-99226ab568be/volumes" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.861216 4867 scope.go:117] "RemoveContainer" containerID="a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.876473 4867 scope.go:117] "RemoveContainer" containerID="6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.896069 4867 scope.go:117] "RemoveContainer" containerID="2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910091 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910198 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z5kv\" (UniqueName: \"kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910329 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910379 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910425 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.910460 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run\") pod \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\" (UID: \"887e1080-2cf6-4c02-a750-b9498fb2f2c0\") " Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911048 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib" (OuterVolumeSpecName: "var-lib") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911107 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log" (OuterVolumeSpecName: "var-log") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911054 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run" (OuterVolumeSpecName: "var-run") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911081 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911478 4867 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-lib\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911510 4867 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-log\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911521 4867 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-var-run\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.911531 4867 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/887e1080-2cf6-4c02-a750-b9498fb2f2c0-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.915173 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv" (OuterVolumeSpecName: "kube-api-access-4z5kv") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "kube-api-access-4z5kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.915344 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts" (OuterVolumeSpecName: "scripts") pod "887e1080-2cf6-4c02-a750-b9498fb2f2c0" (UID: "887e1080-2cf6-4c02-a750-b9498fb2f2c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.926581 4867 scope.go:117] "RemoveContainer" containerID="f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.945371 4867 scope.go:117] "RemoveContainer" containerID="ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.961257 4867 scope.go:117] "RemoveContainer" containerID="09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.961695 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb\": container with ID starting with 09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb not found: ID does not exist" containerID="09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.961731 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb"} err="failed to get container status \"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb\": rpc error: code = NotFound desc = could not find container \"09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb\": container with ID starting with 09f167b75e6adee9739638c85272c9a2f874c5ea3938e83f29688f41c3f69acb not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.961760 4867 scope.go:117] "RemoveContainer" containerID="3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.962051 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070\": container with ID starting with 3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070 not found: ID does not exist" containerID="3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962098 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070"} err="failed to get container status \"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070\": rpc error: code = NotFound desc = could not find container \"3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070\": container with ID starting with 3c964de285be4b53ae7272df9f16b3ebf9ebfd17316aa1b3e138a54b57f3d070 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962129 4867 scope.go:117] "RemoveContainer" containerID="517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.962480 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c\": container with ID starting with 517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c not found: ID does not exist" containerID="517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962508 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c"} err="failed to get container status \"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c\": rpc error: code = NotFound desc = could not find container \"517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c\": container with ID starting with 517a47aa12c391314d57e1bc4ed660b935829066d597bf00576dce6f1a3a441c not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962523 4867 scope.go:117] "RemoveContainer" containerID="b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.962693 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0\": container with ID starting with b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0 not found: ID does not exist" containerID="b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962719 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0"} err="failed to get container status \"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0\": rpc error: code = NotFound desc = could not find container \"b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0\": container with ID starting with b9df326b81875c5663a77dfd04bb90f65571dee09456b15a9334a499937b06f0 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962735 4867 scope.go:117] "RemoveContainer" containerID="b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.962890 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309\": container with ID starting with b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309 not found: ID does not exist" containerID="b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962910 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309"} err="failed to get container status \"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309\": rpc error: code = NotFound desc = could not find container \"b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309\": container with ID starting with b69c3e1b14ee414b31749ce1147b7c368a66e1797ad19a97e9a1b09ed9239309 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.962922 4867 scope.go:117] "RemoveContainer" containerID="6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.963107 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d\": container with ID starting with 6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d not found: ID does not exist" containerID="6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.963257 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d"} err="failed to get container status \"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d\": rpc error: code = NotFound desc = could not find container \"6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d\": container with ID starting with 6ec189384c3c98bb3d9b79602c5cf2b64a674fdb9d53410ec93cebf03d299d6d not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.963323 4867 scope.go:117] "RemoveContainer" containerID="1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.963578 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d\": container with ID starting with 1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d not found: ID does not exist" containerID="1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.963655 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d"} err="failed to get container status \"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d\": rpc error: code = NotFound desc = could not find container \"1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d\": container with ID starting with 1a7ebd09a8f98f239d011122872ba7b1d973aca3d32fb3cd72c5de584ee0048d not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.963720 4867 scope.go:117] "RemoveContainer" containerID="572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.963960 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54\": container with ID starting with 572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54 not found: ID does not exist" containerID="572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.964033 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54"} err="failed to get container status \"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54\": rpc error: code = NotFound desc = could not find container \"572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54\": container with ID starting with 572d030d9ddc65df2d2be85e921d695e3bc5df63c3d76aa9d41cd28045539c54 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.964094 4867 scope.go:117] "RemoveContainer" containerID="bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.964364 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec\": container with ID starting with bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec not found: ID does not exist" containerID="bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.964440 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec"} err="failed to get container status \"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec\": rpc error: code = NotFound desc = could not find container \"bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec\": container with ID starting with bcff99b110f281f1c0c6de2d668d6fff0fa17f0622091af4ae9f63620d7790ec not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.964537 4867 scope.go:117] "RemoveContainer" containerID="6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.964991 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd\": container with ID starting with 6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd not found: ID does not exist" containerID="6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965076 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd"} err="failed to get container status \"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd\": rpc error: code = NotFound desc = could not find container \"6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd\": container with ID starting with 6256460e6862feba1fc0726b5249f0ae6821d159c1eebd8c271000ab063b6bdd not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965150 4867 scope.go:117] "RemoveContainer" containerID="a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.965425 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178\": container with ID starting with a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178 not found: ID does not exist" containerID="a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965504 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178"} err="failed to get container status \"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178\": rpc error: code = NotFound desc = could not find container \"a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178\": container with ID starting with a87062c948f7384597ee10bfc5f3ce42a3df2b51635730424bc1be6f92789178 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965571 4867 scope.go:117] "RemoveContainer" containerID="6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.965946 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751\": container with ID starting with 6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751 not found: ID does not exist" containerID="6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965972 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751"} err="failed to get container status \"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751\": rpc error: code = NotFound desc = could not find container \"6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751\": container with ID starting with 6c554343ef77702e9b7695783c85049e48e28dc9c1837af31b42ecae37159751 not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.965989 4867 scope.go:117] "RemoveContainer" containerID="2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.966464 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc\": container with ID starting with 2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc not found: ID does not exist" containerID="2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.966514 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc"} err="failed to get container status \"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc\": rpc error: code = NotFound desc = could not find container \"2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc\": container with ID starting with 2ce13739d2c9420128032e604f5540cffa2fe60e86de96ef845d52dec8e17ccc not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.966547 4867 scope.go:117] "RemoveContainer" containerID="f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.966998 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb\": container with ID starting with f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb not found: ID does not exist" containerID="f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.967054 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb"} err="failed to get container status \"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb\": rpc error: code = NotFound desc = could not find container \"f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb\": container with ID starting with f8e41e135979734121b212e5d3426389a3b63b1bc1fdca1315df68812f5fcddb not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.967090 4867 scope.go:117] "RemoveContainer" containerID="ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab" Dec 12 07:12:58 crc kubenswrapper[4867]: E1212 07:12:58.967437 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab\": container with ID starting with ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab not found: ID does not exist" containerID="ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.967537 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab"} err="failed to get container status \"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab\": rpc error: code = NotFound desc = could not find container \"ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab\": container with ID starting with ca3a7c6639f0c097918872abc664cfff630f654b5135aab2e9849edb16b24fab not found: ID does not exist" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.988700 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.989002 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.989108 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.989839 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:12:58 crc kubenswrapper[4867]: I1212 07:12:58.989965 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d" gracePeriod=600 Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.012748 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z5kv\" (UniqueName: \"kubernetes.io/projected/887e1080-2cf6-4c02-a750-b9498fb2f2c0-kube-api-access-4z5kv\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.012781 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/887e1080-2cf6-4c02-a750-b9498fb2f2c0-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.667444 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d" exitCode=0 Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.667528 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d"} Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.668039 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda"} Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.668078 4867 scope.go:117] "RemoveContainer" containerID="17662a7f882c56ee4ade45a461826d4fb42ecb6f20133fab8b3d209dfe629d97" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.671262 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9k5zx_887e1080-2cf6-4c02-a750-b9498fb2f2c0/ovs-vswitchd/0.log" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.671980 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9k5zx" event={"ID":"887e1080-2cf6-4c02-a750-b9498fb2f2c0","Type":"ContainerDied","Data":"22e955d75ce13bcaae31e349f84ce73ac23e943d07cc4217976ebd4b9e74c392"} Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.672030 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9k5zx" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.698817 4867 scope.go:117] "RemoveContainer" containerID="fe8abda828a16ee66238fc113149b42592f58fd9e87c826df5eb3dc17509ec11" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.707006 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.714809 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-9k5zx"] Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.717004 4867 scope.go:117] "RemoveContainer" containerID="969ab77bdecad5150743799b08a24c796ea1df712253a8fbd750dde5939590d4" Dec 12 07:12:59 crc kubenswrapper[4867]: I1212 07:12:59.737927 4867 scope.go:117] "RemoveContainer" containerID="f5d8542ed11ac48acd676d04434b7997709cc56b7d071495b3ebe50031d5d89c" Dec 12 07:13:00 crc kubenswrapper[4867]: I1212 07:13:00.853158 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" path="/var/lib/kubelet/pods/887e1080-2cf6-4c02-a750-b9498fb2f2c0/volumes" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.632474 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634241 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38431807-5fbd-4f45-8eca-b42af2c0d221" containerName="keystone-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634275 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="38431807-5fbd-4f45-8eca-b42af2c0d221" containerName="keystone-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634332 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-reaper" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634343 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-reaper" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634362 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634403 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634418 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634427 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634437 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634445 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634488 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634497 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634509 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="ovsdbserver-sb" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634517 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="ovsdbserver-sb" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634525 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="mysql-bootstrap" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634533 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="mysql-bootstrap" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634574 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="mysql-bootstrap" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634582 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="mysql-bootstrap" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.634604 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.634613 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635366 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635381 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635434 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="setup-container" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635442 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="setup-container" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635452 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635459 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635492 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635502 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635520 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635528 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-server" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635571 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635580 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-server" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635591 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-expirer" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635600 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-expirer" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635615 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635645 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635657 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635694 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635747 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635756 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635765 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635802 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635817 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635825 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635834 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="cinder-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635843 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="cinder-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635878 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635886 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635896 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server-init" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635904 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server-init" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635916 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635923 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635952 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635961 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635972 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635981 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.635991 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.635998 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636064 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="swift-recon-cron" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636073 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="swift-recon-cron" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636084 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636090 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636106 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636113 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636162 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636171 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636185 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636193 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636205 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636212 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636239 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-central-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636249 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-central-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636258 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636266 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636279 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636288 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636302 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636309 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636319 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636327 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636338 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636346 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636356 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636363 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636376 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636385 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636398 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636406 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636415 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="dnsmasq-dns" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636422 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="dnsmasq-dns" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636432 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636439 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636449 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="init" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636517 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="init" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636535 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636544 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636555 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437a0623-5c6b-4a29-8da6-f23381fce0ec" containerName="kube-state-metrics" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636563 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="437a0623-5c6b-4a29-8da6-f23381fce0ec" containerName="kube-state-metrics" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636598 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636606 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636620 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636628 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636638 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="sg-core" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636645 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="sg-core" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636679 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="rsync" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636706 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="rsync" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.636717 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.636726 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637061 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637074 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637087 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637094 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637130 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637139 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637152 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8e566d-e439-4e4b-bb80-759e9c4d7f89" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637160 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8e566d-e439-4e4b-bb80-759e9c4d7f89" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637172 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="probe" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637179 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="probe" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637191 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366342dd-c366-4639-9f6d-6444f38fed68" containerName="nova-cell1-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637198 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="366342dd-c366-4639-9f6d-6444f38fed68" containerName="nova-cell1-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637209 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637217 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637244 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637252 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637263 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637270 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-server" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637281 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-notification-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637288 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-notification-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637301 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="setup-container" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637309 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="setup-container" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637321 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637328 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637337 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637345 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637357 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637364 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-server" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637375 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ca9c4-7621-489c-86f2-628a9630039c" containerName="nova-scheduler-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637383 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ca9c4-7621-489c-86f2-628a9630039c" containerName="nova-scheduler-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637395 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637403 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637416 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637422 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637431 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637438 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637446 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45fe5995-c9da-4b48-bef1-20cc51f101f2" containerName="memcached" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637453 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="45fe5995-c9da-4b48-bef1-20cc51f101f2" containerName="memcached" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.637462 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.637470 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638540 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="ovsdbserver-nb" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638577 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="ovsdbserver-nb" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638592 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638600 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638608 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638616 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638652 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638661 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638675 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638682 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638694 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" containerName="nova-cell0-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638701 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" containerName="nova-cell0-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638736 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638745 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-api" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.638759 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.638766 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639131 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f784fec-6124-4aa4-9de3-5d4fd4d58f3f" containerName="nova-cell0-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639146 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="437a0623-5c6b-4a29-8da6-f23381fce0ec" containerName="kube-state-metrics" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639158 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639167 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovsdb-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639179 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639210 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639249 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="ovn-northd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639259 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639269 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="sg-core" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639285 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="00532992-54f5-4a24-b0e4-eb5512e08753" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639297 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-central-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639305 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639339 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="ovsdbserver-nb" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639349 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639364 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7dc7ba4-cc77-4569-ab0f-4bff75ab6472" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639372 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="00269882-fe60-4b26-8d5c-1b34b014b191" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639385 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="887e1080-2cf6-4c02-a750-b9498fb2f2c0" containerName="ovs-vswitchd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639425 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639434 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639445 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639459 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="ovsdbserver-sb" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639466 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="366342dd-c366-4639-9f6d-6444f38fed68" containerName="nova-cell1-conductor-conductor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639503 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639511 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51b1751-7b06-4118-9eff-961ef320bf22" containerName="nova-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639520 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c087ab39-7d5b-42b3-a1f7-5f3703e1d6ba" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639529 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639541 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e267fbd1-d3c8-4379-bb05-e9994fbb13f2" containerName="neutron-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639573 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dac4bdc-46cb-4f7a-9bf3-75566cd8d1c8" containerName="dnsmasq-dns" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639587 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="04569a7e-a52f-4c83-8d2d-e2d6c1ac6d6f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639595 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639607 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d1496c-b4da-404b-9f9c-da2c405070da" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639618 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="45fe5995-c9da-4b48-bef1-20cc51f101f2" containerName="memcached" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639628 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639661 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce609f7-1106-4d01-9a6b-d2099112ce33" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639671 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbb7b7c4-4c08-4f29-8e06-18af9f7422ae" containerName="nova-metadata-metadata" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639681 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8e566d-e439-4e4b-bb80-759e9c4d7f89" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639690 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639704 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a1fd31-a115-4bb6-b65e-70adc5dd77ed" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639740 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639755 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639769 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639781 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639816 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="account-reaper" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639834 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639848 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="38431807-5fbd-4f45-8eca-b42af2c0d221" containerName="keystone-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639855 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="cinder-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639865 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa35f883-b75a-4e67-aeed-f36d00a074cf" containerName="glance-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.639873 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b500049d-056a-40e5-bce8-b6b9f3196d64" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640111 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="ceilometer-notification-agent" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640127 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640139 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640150 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7794fa05-fbb0-4c40-9711-60f299e5ab5b" containerName="barbican-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640161 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="71501691-562e-4385-a19e-fe9f39cfde61" containerName="barbican-worker-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640171 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="163cec12-d32c-4be9-aee0-55703387f2db" containerName="cinder-api" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640184 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640194 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-server" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640203 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d745d-1642-48f4-b1ee-e491f578757e" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640211 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb3e56a-fd7b-4d8e-b42d-2d55de745667" containerName="probe" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640235 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640245 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="container-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640253 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-replicator" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640261 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1b2d98f-1a49-4d66-89b3-ca246c9f227c" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640270 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e39305a-805b-4e9f-ac88-e85e8007409f" containerName="openstack-network-exporter" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640281 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="swift-recon-cron" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640292 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e79e315-4744-4806-96fc-284d97fa88e0" containerName="galera" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640305 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="765da323-2d60-4494-97bc-f699d9bf8661" containerName="placement-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640316 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ad43b4-b2fa-4089-b471-66fbae81fbf2" containerName="rabbitmq" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640326 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-expirer" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640337 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf84c603-ce1e-42be-a595-b27ddc77880c" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640346 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="746ca9c4-7621-489c-86f2-628a9630039c" containerName="nova-scheduler-scheduler" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640354 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-auditor" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640366 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="rsync" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640377 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a11a089-0048-4c70-94e5-9ae18eadfed5" containerName="ovn-controller" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640392 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b1010-d0a8-4c94-a42c-99226ab568be" containerName="object-updater" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640402 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e70b8d-199d-4c5b-be9c-7edc6daf4bc8" containerName="barbican-keystone-listener-log" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640411 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="56091088-518c-4636-bb5e-4bdcc9f0397b" containerName="proxy-httpd" Dec 12 07:13:07 crc kubenswrapper[4867]: E1212 07:13:07.640582 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640591 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f74241-b4d4-41fb-b95f-57534e337198" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.640753 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd36e355-181a-4042-a049-aeda5ec0770f" containerName="mariadb-account-delete" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.641837 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.649530 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.740151 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w6sb\" (UniqueName: \"kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.740483 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.740646 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.842369 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.842477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w6sb\" (UniqueName: \"kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.842513 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.842822 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.842898 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.862825 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w6sb\" (UniqueName: \"kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb\") pod \"redhat-operators-vt9wr\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:07 crc kubenswrapper[4867]: I1212 07:13:07.958346 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:08 crc kubenswrapper[4867]: I1212 07:13:08.400268 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:08 crc kubenswrapper[4867]: I1212 07:13:08.762950 4867 generic.go:334] "Generic (PLEG): container finished" podID="a922815f-d068-4100-8d73-ed88926500a2" containerID="5a77cacedbc687d041a6c27433cd7f750c2cb6f8f74e44f0b6cb78d911f2ec9b" exitCode=0 Dec 12 07:13:08 crc kubenswrapper[4867]: I1212 07:13:08.763003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerDied","Data":"5a77cacedbc687d041a6c27433cd7f750c2cb6f8f74e44f0b6cb78d911f2ec9b"} Dec 12 07:13:08 crc kubenswrapper[4867]: I1212 07:13:08.763287 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerStarted","Data":"f45d38a7550801b2f24bda9b506956fa96ed2c65e65e63f1f775762dcba3cd49"} Dec 12 07:13:10 crc kubenswrapper[4867]: I1212 07:13:10.782075 4867 generic.go:334] "Generic (PLEG): container finished" podID="a922815f-d068-4100-8d73-ed88926500a2" containerID="af01664291de0860a0984bb56bbaf208d5866bd10ca8ccdc565941de63bcbe72" exitCode=0 Dec 12 07:13:10 crc kubenswrapper[4867]: I1212 07:13:10.782113 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerDied","Data":"af01664291de0860a0984bb56bbaf208d5866bd10ca8ccdc565941de63bcbe72"} Dec 12 07:13:12 crc kubenswrapper[4867]: I1212 07:13:12.800502 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerStarted","Data":"ab5cb364c20b3a5658f2540f1295ca643e9c2f51911b026b8e69cdf484d8d77b"} Dec 12 07:13:12 crc kubenswrapper[4867]: I1212 07:13:12.821031 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vt9wr" podStartSLOduration=2.965426296 podStartE2EDuration="5.821014448s" podCreationTimestamp="2025-12-12 07:13:07 +0000 UTC" firstStartedPulling="2025-12-12 07:13:08.774694699 +0000 UTC m=+1476.346075968" lastFinishedPulling="2025-12-12 07:13:11.630282851 +0000 UTC m=+1479.201664120" observedRunningTime="2025-12-12 07:13:12.816778151 +0000 UTC m=+1480.388159430" watchObservedRunningTime="2025-12-12 07:13:12.821014448 +0000 UTC m=+1480.392395717" Dec 12 07:13:17 crc kubenswrapper[4867]: I1212 07:13:17.958927 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:17 crc kubenswrapper[4867]: I1212 07:13:17.959346 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:18 crc kubenswrapper[4867]: I1212 07:13:18.005480 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:18 crc kubenswrapper[4867]: I1212 07:13:18.900979 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:18 crc kubenswrapper[4867]: I1212 07:13:18.947946 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:20 crc kubenswrapper[4867]: I1212 07:13:20.872751 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vt9wr" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="registry-server" containerID="cri-o://ab5cb364c20b3a5658f2540f1295ca643e9c2f51911b026b8e69cdf484d8d77b" gracePeriod=2 Dec 12 07:13:22 crc kubenswrapper[4867]: I1212 07:13:22.898639 4867 generic.go:334] "Generic (PLEG): container finished" podID="a922815f-d068-4100-8d73-ed88926500a2" containerID="ab5cb364c20b3a5658f2540f1295ca643e9c2f51911b026b8e69cdf484d8d77b" exitCode=0 Dec 12 07:13:22 crc kubenswrapper[4867]: I1212 07:13:22.898717 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerDied","Data":"ab5cb364c20b3a5658f2540f1295ca643e9c2f51911b026b8e69cdf484d8d77b"} Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.063609 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.155389 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities\") pod \"a922815f-d068-4100-8d73-ed88926500a2\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.155511 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content\") pod \"a922815f-d068-4100-8d73-ed88926500a2\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.155587 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w6sb\" (UniqueName: \"kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb\") pod \"a922815f-d068-4100-8d73-ed88926500a2\" (UID: \"a922815f-d068-4100-8d73-ed88926500a2\") " Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.156509 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities" (OuterVolumeSpecName: "utilities") pod "a922815f-d068-4100-8d73-ed88926500a2" (UID: "a922815f-d068-4100-8d73-ed88926500a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.161294 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb" (OuterVolumeSpecName: "kube-api-access-7w6sb") pod "a922815f-d068-4100-8d73-ed88926500a2" (UID: "a922815f-d068-4100-8d73-ed88926500a2"). InnerVolumeSpecName "kube-api-access-7w6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.257174 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.257265 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w6sb\" (UniqueName: \"kubernetes.io/projected/a922815f-d068-4100-8d73-ed88926500a2-kube-api-access-7w6sb\") on node \"crc\" DevicePath \"\"" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.274206 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a922815f-d068-4100-8d73-ed88926500a2" (UID: "a922815f-d068-4100-8d73-ed88926500a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.358542 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a922815f-d068-4100-8d73-ed88926500a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.920656 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vt9wr" event={"ID":"a922815f-d068-4100-8d73-ed88926500a2","Type":"ContainerDied","Data":"f45d38a7550801b2f24bda9b506956fa96ed2c65e65e63f1f775762dcba3cd49"} Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.920791 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vt9wr" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.921003 4867 scope.go:117] "RemoveContainer" containerID="ab5cb364c20b3a5658f2540f1295ca643e9c2f51911b026b8e69cdf484d8d77b" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.951673 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.953303 4867 scope.go:117] "RemoveContainer" containerID="af01664291de0860a0984bb56bbaf208d5866bd10ca8ccdc565941de63bcbe72" Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.957169 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vt9wr"] Dec 12 07:13:23 crc kubenswrapper[4867]: I1212 07:13:23.976873 4867 scope.go:117] "RemoveContainer" containerID="5a77cacedbc687d041a6c27433cd7f750c2cb6f8f74e44f0b6cb78d911f2ec9b" Dec 12 07:13:24 crc kubenswrapper[4867]: I1212 07:13:24.847660 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a922815f-d068-4100-8d73-ed88926500a2" path="/var/lib/kubelet/pods/a922815f-d068-4100-8d73-ed88926500a2/volumes" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.525599 4867 scope.go:117] "RemoveContainer" containerID="e951181ca2240c8345911e4e0b3c224a258ab5437f10f82e6d47edbc589d9fb2" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.557263 4867 scope.go:117] "RemoveContainer" containerID="651d2a085bb2c2fbadc12806221bb243108cd7ac6a3dcb3f1154eed3ff0a88ed" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.591385 4867 scope.go:117] "RemoveContainer" containerID="62cb32eb7c19228623a0cc83eac79f7834312e294015b665c1c83d3000645c7f" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.632568 4867 scope.go:117] "RemoveContainer" containerID="c5dab6101821935c5fac66af5d8442253153302d4354001cf035fc6582565ef1" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.657153 4867 scope.go:117] "RemoveContainer" containerID="04e9eda5d1420cdeac04a51995e5de475c02ac1d5311c054f1124ae87b3fb8f0" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.677626 4867 scope.go:117] "RemoveContainer" containerID="8c81c50804367f392fba39b977b26bfecdb90e8e5eb09c175d3b03b46dbc1469" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.703585 4867 scope.go:117] "RemoveContainer" containerID="fe84ae831036079376416809cc68115d051b2918c176b712ed73dbcf926065ad" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.773571 4867 scope.go:117] "RemoveContainer" containerID="61dac5c9e6307be96fa44d7a6e3d7fc40e109ecc80d7a7042569d594f2df0414" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.799825 4867 scope.go:117] "RemoveContainer" containerID="77646526d618a77ee8b6ba9c19111bdb9e0e386abe004e3170ce8918fdae8267" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.852457 4867 scope.go:117] "RemoveContainer" containerID="4577bfb6bb31cfe2a93939f5385f788d8aa1c3bb19025613cf01defc00d6a24d" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.884770 4867 scope.go:117] "RemoveContainer" containerID="ebaea2e9bb46bdb9b9f5d3e785667e43809d4ca7be867bac6014e24edd424184" Dec 12 07:13:40 crc kubenswrapper[4867]: I1212 07:13:40.955283 4867 scope.go:117] "RemoveContainer" containerID="dea0997833a5437fd02115c35913afaf141ffd915e7a817d66c7a6536e56241f" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.002165 4867 scope.go:117] "RemoveContainer" containerID="7a47e43be70ffc5bba10ae4e7296ef0e5296adf16d60fc3a9122ae9a7df19a71" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.027384 4867 scope.go:117] "RemoveContainer" containerID="01bac50358ef03d8b64fd11175d1911fa871e096ed6fc433023c26d56989cc80" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.047925 4867 scope.go:117] "RemoveContainer" containerID="30179a07ae8844a4ffe7aef6f4c96372c3ae3317a61f76efde0b33fbe5dc214c" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.065173 4867 scope.go:117] "RemoveContainer" containerID="bcf8eb7e4e5ab4fec8f5c52187980849509ed93d855f00433e7a743ea90ffb05" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.117888 4867 scope.go:117] "RemoveContainer" containerID="77c9a26c17603a8e3b6947d790bbaaaac80e35891cbed056027c6b7726af2e06" Dec 12 07:13:41 crc kubenswrapper[4867]: I1212 07:13:41.155768 4867 scope.go:117] "RemoveContainer" containerID="8e0865343bc535ead2fc7e07bf96597f4556a3b0218b2ab297f2420213bffa0f" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.064312 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:27 crc kubenswrapper[4867]: E1212 07:14:27.065105 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="extract-utilities" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.065120 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="extract-utilities" Dec 12 07:14:27 crc kubenswrapper[4867]: E1212 07:14:27.065139 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="registry-server" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.065146 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="registry-server" Dec 12 07:14:27 crc kubenswrapper[4867]: E1212 07:14:27.065171 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="extract-content" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.065179 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="extract-content" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.065352 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a922815f-d068-4100-8d73-ed88926500a2" containerName="registry-server" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.066347 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.078578 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.221941 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.221991 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.222019 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whsf9\" (UniqueName: \"kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.323597 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.323655 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.323696 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whsf9\" (UniqueName: \"kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.324202 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.324493 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.347384 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whsf9\" (UniqueName: \"kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9\") pod \"community-operators-n67cc\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.383988 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:27 crc kubenswrapper[4867]: I1212 07:14:27.722135 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:28 crc kubenswrapper[4867]: I1212 07:14:28.458253 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerStarted","Data":"4a396731379e8d6a264587023e0b1b0e3b269888281e65442895c13f3eb8f4c5"} Dec 12 07:14:29 crc kubenswrapper[4867]: I1212 07:14:29.474251 4867 generic.go:334] "Generic (PLEG): container finished" podID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerID="d5a1ae853494d43bbe34992e4201c3c28432b13d2e89c9f883da1e2c1061dfd0" exitCode=0 Dec 12 07:14:29 crc kubenswrapper[4867]: I1212 07:14:29.474337 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerDied","Data":"d5a1ae853494d43bbe34992e4201c3c28432b13d2e89c9f883da1e2c1061dfd0"} Dec 12 07:14:30 crc kubenswrapper[4867]: E1212 07:14:30.965809 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod171fc78b_91fa_42e2_8351_dee61ebf8648.slice/crio-conmon-696efd242132c3b279ed3ce2275d19c8bf8fa9fbd2b49667be8ee4a8148c6f61.scope\": RecentStats: unable to find data in memory cache]" Dec 12 07:14:31 crc kubenswrapper[4867]: I1212 07:14:31.494307 4867 generic.go:334] "Generic (PLEG): container finished" podID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerID="696efd242132c3b279ed3ce2275d19c8bf8fa9fbd2b49667be8ee4a8148c6f61" exitCode=0 Dec 12 07:14:31 crc kubenswrapper[4867]: I1212 07:14:31.494367 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerDied","Data":"696efd242132c3b279ed3ce2275d19c8bf8fa9fbd2b49667be8ee4a8148c6f61"} Dec 12 07:14:32 crc kubenswrapper[4867]: I1212 07:14:32.504331 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerStarted","Data":"f77637abe1b96cf623eb39e2b524f9ca3810f21996332e688f602bf7ec592e8b"} Dec 12 07:14:32 crc kubenswrapper[4867]: I1212 07:14:32.522632 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n67cc" podStartSLOduration=2.953973637 podStartE2EDuration="5.522610162s" podCreationTimestamp="2025-12-12 07:14:27 +0000 UTC" firstStartedPulling="2025-12-12 07:14:29.476612683 +0000 UTC m=+1557.047993952" lastFinishedPulling="2025-12-12 07:14:32.045249208 +0000 UTC m=+1559.616630477" observedRunningTime="2025-12-12 07:14:32.520165112 +0000 UTC m=+1560.091546381" watchObservedRunningTime="2025-12-12 07:14:32.522610162 +0000 UTC m=+1560.093991431" Dec 12 07:14:37 crc kubenswrapper[4867]: I1212 07:14:37.384620 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:37 crc kubenswrapper[4867]: I1212 07:14:37.385510 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:37 crc kubenswrapper[4867]: I1212 07:14:37.442213 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:37 crc kubenswrapper[4867]: I1212 07:14:37.604183 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:37 crc kubenswrapper[4867]: I1212 07:14:37.679893 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:39 crc kubenswrapper[4867]: I1212 07:14:39.574074 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n67cc" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="registry-server" containerID="cri-o://f77637abe1b96cf623eb39e2b524f9ca3810f21996332e688f602bf7ec592e8b" gracePeriod=2 Dec 12 07:14:41 crc kubenswrapper[4867]: I1212 07:14:41.772884 4867 scope.go:117] "RemoveContainer" containerID="3f67093be35c467d200dbddfac2e3e8c6dc0d42da9b5d84d7d84b0c2b0ca891d" Dec 12 07:14:41 crc kubenswrapper[4867]: I1212 07:14:41.821325 4867 scope.go:117] "RemoveContainer" containerID="fb810476f4594e24532a01d472f2d4a6803aa3cc6fe3bdf084703be4a759b4ce" Dec 12 07:14:41 crc kubenswrapper[4867]: I1212 07:14:41.857152 4867 scope.go:117] "RemoveContainer" containerID="38a25c4bf9c0c0aa9898ef57ec5198b23b894a9894b5b11dd705553353f23d8d" Dec 12 07:14:41 crc kubenswrapper[4867]: I1212 07:14:41.889395 4867 scope.go:117] "RemoveContainer" containerID="ba666498132b53c649f4905c4a572219fb0ab53120aad5d06225092d208e3c0d" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.617521 4867 generic.go:334] "Generic (PLEG): container finished" podID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerID="f77637abe1b96cf623eb39e2b524f9ca3810f21996332e688f602bf7ec592e8b" exitCode=0 Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.617566 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerDied","Data":"f77637abe1b96cf623eb39e2b524f9ca3810f21996332e688f602bf7ec592e8b"} Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.643772 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.750129 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whsf9\" (UniqueName: \"kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9\") pod \"171fc78b-91fa-42e2-8351-dee61ebf8648\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.750622 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities\") pod \"171fc78b-91fa-42e2-8351-dee61ebf8648\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.750722 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content\") pod \"171fc78b-91fa-42e2-8351-dee61ebf8648\" (UID: \"171fc78b-91fa-42e2-8351-dee61ebf8648\") " Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.751345 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities" (OuterVolumeSpecName: "utilities") pod "171fc78b-91fa-42e2-8351-dee61ebf8648" (UID: "171fc78b-91fa-42e2-8351-dee61ebf8648"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.754913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9" (OuterVolumeSpecName: "kube-api-access-whsf9") pod "171fc78b-91fa-42e2-8351-dee61ebf8648" (UID: "171fc78b-91fa-42e2-8351-dee61ebf8648"). InnerVolumeSpecName "kube-api-access-whsf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.803781 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "171fc78b-91fa-42e2-8351-dee61ebf8648" (UID: "171fc78b-91fa-42e2-8351-dee61ebf8648"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.852030 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.852072 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whsf9\" (UniqueName: \"kubernetes.io/projected/171fc78b-91fa-42e2-8351-dee61ebf8648-kube-api-access-whsf9\") on node \"crc\" DevicePath \"\"" Dec 12 07:14:42 crc kubenswrapper[4867]: I1212 07:14:42.852087 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/171fc78b-91fa-42e2-8351-dee61ebf8648-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.627486 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n67cc" event={"ID":"171fc78b-91fa-42e2-8351-dee61ebf8648","Type":"ContainerDied","Data":"4a396731379e8d6a264587023e0b1b0e3b269888281e65442895c13f3eb8f4c5"} Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.627543 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n67cc" Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.627603 4867 scope.go:117] "RemoveContainer" containerID="f77637abe1b96cf623eb39e2b524f9ca3810f21996332e688f602bf7ec592e8b" Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.645417 4867 scope.go:117] "RemoveContainer" containerID="696efd242132c3b279ed3ce2275d19c8bf8fa9fbd2b49667be8ee4a8148c6f61" Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.664027 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.669493 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n67cc"] Dec 12 07:14:43 crc kubenswrapper[4867]: I1212 07:14:43.671804 4867 scope.go:117] "RemoveContainer" containerID="d5a1ae853494d43bbe34992e4201c3c28432b13d2e89c9f883da1e2c1061dfd0" Dec 12 07:14:44 crc kubenswrapper[4867]: I1212 07:14:44.847984 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" path="/var/lib/kubelet/pods/171fc78b-91fa-42e2-8351-dee61ebf8648/volumes" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.139164 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59"] Dec 12 07:15:00 crc kubenswrapper[4867]: E1212 07:15:00.140127 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="registry-server" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.140145 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="registry-server" Dec 12 07:15:00 crc kubenswrapper[4867]: E1212 07:15:00.140164 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="extract-content" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.140172 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="extract-content" Dec 12 07:15:00 crc kubenswrapper[4867]: E1212 07:15:00.140185 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="extract-utilities" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.140193 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="extract-utilities" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.143429 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="171fc78b-91fa-42e2-8351-dee61ebf8648" containerName="registry-server" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.144447 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.148426 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.148708 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.167488 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59"] Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.215261 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.215347 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.215379 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cfrp\" (UniqueName: \"kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.317279 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cfrp\" (UniqueName: \"kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.317414 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.317471 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.318674 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.326282 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.332835 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cfrp\" (UniqueName: \"kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp\") pod \"collect-profiles-29425395-qbb59\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.469720 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:00 crc kubenswrapper[4867]: I1212 07:15:00.888975 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59"] Dec 12 07:15:01 crc kubenswrapper[4867]: E1212 07:15:01.532029 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9f4224a_94dd_44b0_a236_d329322edbf2.slice/crio-263906e1ec399971bc56c4340068b099c334fea081d41bee2129d19a6de510c4.scope\": RecentStats: unable to find data in memory cache]" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.610300 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.617660 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.620823 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.636460 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxlcs\" (UniqueName: \"kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.636509 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.636612 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.737623 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxlcs\" (UniqueName: \"kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.737668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.737767 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.738377 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.738673 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.757611 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxlcs\" (UniqueName: \"kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs\") pod \"certified-operators-fkgcc\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.816752 4867 generic.go:334] "Generic (PLEG): container finished" podID="d9f4224a-94dd-44b0-a236-d329322edbf2" containerID="263906e1ec399971bc56c4340068b099c334fea081d41bee2129d19a6de510c4" exitCode=0 Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.816796 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" event={"ID":"d9f4224a-94dd-44b0-a236-d329322edbf2","Type":"ContainerDied","Data":"263906e1ec399971bc56c4340068b099c334fea081d41bee2129d19a6de510c4"} Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.816823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" event={"ID":"d9f4224a-94dd-44b0-a236-d329322edbf2","Type":"ContainerStarted","Data":"034d260360c07b59246c1a1fc6d7327aafcfb8a9b249d8fb8f72f6679f9917eb"} Dec 12 07:15:01 crc kubenswrapper[4867]: I1212 07:15:01.939949 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:02 crc kubenswrapper[4867]: W1212 07:15:02.482330 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9b4b9ca_ef06_4750_a91a_39c903cc6e65.slice/crio-5518f9fcd8db119430ed39b3d9510a7e1d32cc81c091ef5560962ffb8d209ac3 WatchSource:0}: Error finding container 5518f9fcd8db119430ed39b3d9510a7e1d32cc81c091ef5560962ffb8d209ac3: Status 404 returned error can't find the container with id 5518f9fcd8db119430ed39b3d9510a7e1d32cc81c091ef5560962ffb8d209ac3 Dec 12 07:15:02 crc kubenswrapper[4867]: I1212 07:15:02.486926 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:02 crc kubenswrapper[4867]: I1212 07:15:02.833321 4867 generic.go:334] "Generic (PLEG): container finished" podID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerID="9c8c791d7083e2102204a41a99f6b55375a9a53ff2f9d33c917b19a38170f602" exitCode=0 Dec 12 07:15:02 crc kubenswrapper[4867]: I1212 07:15:02.833915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerDied","Data":"9c8c791d7083e2102204a41a99f6b55375a9a53ff2f9d33c917b19a38170f602"} Dec 12 07:15:02 crc kubenswrapper[4867]: I1212 07:15:02.833970 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerStarted","Data":"5518f9fcd8db119430ed39b3d9510a7e1d32cc81c091ef5560962ffb8d209ac3"} Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.137421 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.206802 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume\") pod \"d9f4224a-94dd-44b0-a236-d329322edbf2\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.206881 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume\") pod \"d9f4224a-94dd-44b0-a236-d329322edbf2\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.207025 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cfrp\" (UniqueName: \"kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp\") pod \"d9f4224a-94dd-44b0-a236-d329322edbf2\" (UID: \"d9f4224a-94dd-44b0-a236-d329322edbf2\") " Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.217325 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume" (OuterVolumeSpecName: "config-volume") pod "d9f4224a-94dd-44b0-a236-d329322edbf2" (UID: "d9f4224a-94dd-44b0-a236-d329322edbf2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.242095 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d9f4224a-94dd-44b0-a236-d329322edbf2" (UID: "d9f4224a-94dd-44b0-a236-d329322edbf2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.242424 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp" (OuterVolumeSpecName: "kube-api-access-7cfrp") pod "d9f4224a-94dd-44b0-a236-d329322edbf2" (UID: "d9f4224a-94dd-44b0-a236-d329322edbf2"). InnerVolumeSpecName "kube-api-access-7cfrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.309890 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d9f4224a-94dd-44b0-a236-d329322edbf2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.309951 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d9f4224a-94dd-44b0-a236-d329322edbf2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.309961 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cfrp\" (UniqueName: \"kubernetes.io/projected/d9f4224a-94dd-44b0-a236-d329322edbf2-kube-api-access-7cfrp\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.841853 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" event={"ID":"d9f4224a-94dd-44b0-a236-d329322edbf2","Type":"ContainerDied","Data":"034d260360c07b59246c1a1fc6d7327aafcfb8a9b249d8fb8f72f6679f9917eb"} Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.842715 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="034d260360c07b59246c1a1fc6d7327aafcfb8a9b249d8fb8f72f6679f9917eb" Dec 12 07:15:03 crc kubenswrapper[4867]: I1212 07:15:03.841920 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59" Dec 12 07:15:04 crc kubenswrapper[4867]: I1212 07:15:04.854065 4867 generic.go:334] "Generic (PLEG): container finished" podID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerID="12dce599206cb6dfcf88ebd49e5c2da3a041fa0d154e1ef21b3c7cb49629908e" exitCode=0 Dec 12 07:15:04 crc kubenswrapper[4867]: I1212 07:15:04.854127 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerDied","Data":"12dce599206cb6dfcf88ebd49e5c2da3a041fa0d154e1ef21b3c7cb49629908e"} Dec 12 07:15:06 crc kubenswrapper[4867]: I1212 07:15:06.873465 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerStarted","Data":"b79e82b69b33a2c7f70628619857b29af03e60cf80c58b395dd7dfa77c58bffe"} Dec 12 07:15:11 crc kubenswrapper[4867]: I1212 07:15:11.940260 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:11 crc kubenswrapper[4867]: I1212 07:15:11.941655 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:11 crc kubenswrapper[4867]: I1212 07:15:11.982677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:12 crc kubenswrapper[4867]: I1212 07:15:11.999995 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fkgcc" podStartSLOduration=8.058231217 podStartE2EDuration="10.999976526s" podCreationTimestamp="2025-12-12 07:15:01 +0000 UTC" firstStartedPulling="2025-12-12 07:15:02.852362286 +0000 UTC m=+1590.423743555" lastFinishedPulling="2025-12-12 07:15:05.794107595 +0000 UTC m=+1593.365488864" observedRunningTime="2025-12-12 07:15:06.898049604 +0000 UTC m=+1594.469430873" watchObservedRunningTime="2025-12-12 07:15:11.999976526 +0000 UTC m=+1599.571357785" Dec 12 07:15:13 crc kubenswrapper[4867]: I1212 07:15:13.007603 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:13 crc kubenswrapper[4867]: I1212 07:15:13.062529 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:14 crc kubenswrapper[4867]: I1212 07:15:14.953121 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fkgcc" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="registry-server" containerID="cri-o://b79e82b69b33a2c7f70628619857b29af03e60cf80c58b395dd7dfa77c58bffe" gracePeriod=2 Dec 12 07:15:16 crc kubenswrapper[4867]: I1212 07:15:16.973790 4867 generic.go:334] "Generic (PLEG): container finished" podID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerID="b79e82b69b33a2c7f70628619857b29af03e60cf80c58b395dd7dfa77c58bffe" exitCode=0 Dec 12 07:15:16 crc kubenswrapper[4867]: I1212 07:15:16.973897 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerDied","Data":"b79e82b69b33a2c7f70628619857b29af03e60cf80c58b395dd7dfa77c58bffe"} Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.160458 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.214948 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities\") pod \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.215076 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxlcs\" (UniqueName: \"kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs\") pod \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.215148 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content\") pod \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\" (UID: \"f9b4b9ca-ef06-4750-a91a-39c903cc6e65\") " Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.216063 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities" (OuterVolumeSpecName: "utilities") pod "f9b4b9ca-ef06-4750-a91a-39c903cc6e65" (UID: "f9b4b9ca-ef06-4750-a91a-39c903cc6e65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.220970 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs" (OuterVolumeSpecName: "kube-api-access-bxlcs") pod "f9b4b9ca-ef06-4750-a91a-39c903cc6e65" (UID: "f9b4b9ca-ef06-4750-a91a-39c903cc6e65"). InnerVolumeSpecName "kube-api-access-bxlcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.275027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9b4b9ca-ef06-4750-a91a-39c903cc6e65" (UID: "f9b4b9ca-ef06-4750-a91a-39c903cc6e65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.316734 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxlcs\" (UniqueName: \"kubernetes.io/projected/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-kube-api-access-bxlcs\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.316777 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.316787 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9b4b9ca-ef06-4750-a91a-39c903cc6e65-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.986629 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fkgcc" event={"ID":"f9b4b9ca-ef06-4750-a91a-39c903cc6e65","Type":"ContainerDied","Data":"5518f9fcd8db119430ed39b3d9510a7e1d32cc81c091ef5560962ffb8d209ac3"} Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.986694 4867 scope.go:117] "RemoveContainer" containerID="b79e82b69b33a2c7f70628619857b29af03e60cf80c58b395dd7dfa77c58bffe" Dec 12 07:15:17 crc kubenswrapper[4867]: I1212 07:15:17.986857 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fkgcc" Dec 12 07:15:18 crc kubenswrapper[4867]: I1212 07:15:18.008183 4867 scope.go:117] "RemoveContainer" containerID="12dce599206cb6dfcf88ebd49e5c2da3a041fa0d154e1ef21b3c7cb49629908e" Dec 12 07:15:18 crc kubenswrapper[4867]: I1212 07:15:18.019148 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:18 crc kubenswrapper[4867]: I1212 07:15:18.024508 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fkgcc"] Dec 12 07:15:18 crc kubenswrapper[4867]: I1212 07:15:18.048667 4867 scope.go:117] "RemoveContainer" containerID="9c8c791d7083e2102204a41a99f6b55375a9a53ff2f9d33c917b19a38170f602" Dec 12 07:15:18 crc kubenswrapper[4867]: I1212 07:15:18.846291 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" path="/var/lib/kubelet/pods/f9b4b9ca-ef06-4750-a91a-39c903cc6e65/volumes" Dec 12 07:15:28 crc kubenswrapper[4867]: I1212 07:15:28.988583 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:15:28 crc kubenswrapper[4867]: I1212 07:15:28.989176 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:15:41 crc kubenswrapper[4867]: I1212 07:15:41.980450 4867 scope.go:117] "RemoveContainer" containerID="2c02009ab1e1505ffea6c57acadcdd4d133ceb5623c0d8e451ec9680ea621aeb" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.034072 4867 scope.go:117] "RemoveContainer" containerID="c73943457de5fe6564da42074030b9d5a1e4173779760522de4f412b52e3c63e" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.055598 4867 scope.go:117] "RemoveContainer" containerID="8cc586c884d2d16c09c77de72811d41b3e24fe24ed8d99ac084d0cfb59b68ee4" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.079163 4867 scope.go:117] "RemoveContainer" containerID="97739a03e3bc94aca021895c2059a73f840207d78f95cbff068baa5aa4b4baf0" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.102173 4867 scope.go:117] "RemoveContainer" containerID="fbc9e2de34a9d4fa8a9c7a372b1fa0f7bd6ae42ab2cbe3e5311d425d777dafd7" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.127253 4867 scope.go:117] "RemoveContainer" containerID="fd75e02b1ef268a12725da9e2896abd47d835148c1ca4dace74da4dd77e5455d" Dec 12 07:15:42 crc kubenswrapper[4867]: I1212 07:15:42.153828 4867 scope.go:117] "RemoveContainer" containerID="f052721f144c8d81984067d872d7c79d7f1cda0b9aede5c27429be698c0098d5" Dec 12 07:15:58 crc kubenswrapper[4867]: I1212 07:15:58.988593 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:15:58 crc kubenswrapper[4867]: I1212 07:15:58.989206 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:16:28 crc kubenswrapper[4867]: I1212 07:16:28.989255 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:16:28 crc kubenswrapper[4867]: I1212 07:16:28.989833 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:16:28 crc kubenswrapper[4867]: I1212 07:16:28.989880 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:16:28 crc kubenswrapper[4867]: I1212 07:16:28.990523 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:16:28 crc kubenswrapper[4867]: I1212 07:16:28.990589 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" gracePeriod=600 Dec 12 07:16:29 crc kubenswrapper[4867]: I1212 07:16:29.537566 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" exitCode=0 Dec 12 07:16:29 crc kubenswrapper[4867]: I1212 07:16:29.537607 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda"} Dec 12 07:16:29 crc kubenswrapper[4867]: I1212 07:16:29.538019 4867 scope.go:117] "RemoveContainer" containerID="e75d77a5f834722065a1ea0f8d46d51d3f62d40bb2c042d206aefae5baab539d" Dec 12 07:16:30 crc kubenswrapper[4867]: E1212 07:16:30.003431 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:16:30 crc kubenswrapper[4867]: I1212 07:16:30.548436 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:16:30 crc kubenswrapper[4867]: E1212 07:16:30.550382 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:16:42 crc kubenswrapper[4867]: I1212 07:16:42.280042 4867 scope.go:117] "RemoveContainer" containerID="dc7c6abeff19f2ed7d1082efbb635045ee7b036f7c1606323782742b6b4993b6" Dec 12 07:16:42 crc kubenswrapper[4867]: I1212 07:16:42.320781 4867 scope.go:117] "RemoveContainer" containerID="eacc7194f20858f2dea819114f253d015cd6011df2db04d13f28c44e7fb5a85f" Dec 12 07:16:42 crc kubenswrapper[4867]: I1212 07:16:42.342114 4867 scope.go:117] "RemoveContainer" containerID="1671dbf245b4faa8dcbc3ca3be38cf020c46644501560453930a7a1ee4255c29" Dec 12 07:16:42 crc kubenswrapper[4867]: I1212 07:16:42.359100 4867 scope.go:117] "RemoveContainer" containerID="1eb349ad10dac75fbf2979864d21a0ab5fd4b851e81f9bfe08f8809cf9cf72f8" Dec 12 07:16:45 crc kubenswrapper[4867]: I1212 07:16:45.838626 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:16:45 crc kubenswrapper[4867]: E1212 07:16:45.839396 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:16:57 crc kubenswrapper[4867]: I1212 07:16:57.838320 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:16:57 crc kubenswrapper[4867]: E1212 07:16:57.839127 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:17:09 crc kubenswrapper[4867]: I1212 07:17:09.838335 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:17:09 crc kubenswrapper[4867]: E1212 07:17:09.839041 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:17:21 crc kubenswrapper[4867]: I1212 07:17:21.838433 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:17:21 crc kubenswrapper[4867]: E1212 07:17:21.839295 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:17:32 crc kubenswrapper[4867]: I1212 07:17:32.841942 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:17:32 crc kubenswrapper[4867]: E1212 07:17:32.843102 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:17:42 crc kubenswrapper[4867]: I1212 07:17:42.436778 4867 scope.go:117] "RemoveContainer" containerID="334aa3a251311e5ad85b964132678178f0417b43878df9f044e9cc662fc15fbd" Dec 12 07:17:42 crc kubenswrapper[4867]: I1212 07:17:42.458277 4867 scope.go:117] "RemoveContainer" containerID="5b998256ed7e715606082bd623b651120a069bd1bdab372dad3314ef077aeaaf" Dec 12 07:17:42 crc kubenswrapper[4867]: I1212 07:17:42.481479 4867 scope.go:117] "RemoveContainer" containerID="36cf7873d0b7a0123d07b6351629b53917faf6e82674638ceaca8e46d8aff9e8" Dec 12 07:17:42 crc kubenswrapper[4867]: I1212 07:17:42.499277 4867 scope.go:117] "RemoveContainer" containerID="a0bedba8914b8a45bc1c0ddce256133a717fdad864ae53b01a0eb5e13c53466b" Dec 12 07:17:42 crc kubenswrapper[4867]: I1212 07:17:42.513363 4867 scope.go:117] "RemoveContainer" containerID="ed60bf6d99ceafbd9fedf17a5214f14b39f75b17d2af05f1022ba59e01bf9864" Dec 12 07:17:45 crc kubenswrapper[4867]: I1212 07:17:45.838308 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:17:45 crc kubenswrapper[4867]: E1212 07:17:45.838805 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:17:56 crc kubenswrapper[4867]: I1212 07:17:56.838131 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:17:56 crc kubenswrapper[4867]: E1212 07:17:56.839062 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:18:11 crc kubenswrapper[4867]: I1212 07:18:11.838602 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:18:11 crc kubenswrapper[4867]: E1212 07:18:11.840128 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.064475 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:17 crc kubenswrapper[4867]: E1212 07:18:17.069541 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="extract-utilities" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069566 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="extract-utilities" Dec 12 07:18:17 crc kubenswrapper[4867]: E1212 07:18:17.069585 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f4224a-94dd-44b0-a236-d329322edbf2" containerName="collect-profiles" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069593 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f4224a-94dd-44b0-a236-d329322edbf2" containerName="collect-profiles" Dec 12 07:18:17 crc kubenswrapper[4867]: E1212 07:18:17.069607 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="extract-content" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069616 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="extract-content" Dec 12 07:18:17 crc kubenswrapper[4867]: E1212 07:18:17.069638 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="registry-server" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069649 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="registry-server" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069819 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f4224a-94dd-44b0-a236-d329322edbf2" containerName="collect-profiles" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.069837 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9b4b9ca-ef06-4750-a91a-39c903cc6e65" containerName="registry-server" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.071016 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.073983 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.123130 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqljv\" (UniqueName: \"kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.123273 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.123315 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.224257 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.224336 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.224389 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqljv\" (UniqueName: \"kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.224735 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.224902 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.245706 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqljv\" (UniqueName: \"kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv\") pod \"redhat-marketplace-wd9dl\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.394440 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:17 crc kubenswrapper[4867]: I1212 07:18:17.624498 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:18 crc kubenswrapper[4867]: I1212 07:18:18.341065 4867 generic.go:334] "Generic (PLEG): container finished" podID="18d7947e-f766-4149-9e15-ce48a705d4de" containerID="ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33" exitCode=0 Dec 12 07:18:18 crc kubenswrapper[4867]: I1212 07:18:18.341124 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerDied","Data":"ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33"} Dec 12 07:18:18 crc kubenswrapper[4867]: I1212 07:18:18.341403 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerStarted","Data":"486ede1f6083f3d1d87835bbf08489ef49208a5ec6fd45445b4b93cd413356f0"} Dec 12 07:18:18 crc kubenswrapper[4867]: I1212 07:18:18.342780 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:18:19 crc kubenswrapper[4867]: I1212 07:18:19.350970 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerStarted","Data":"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912"} Dec 12 07:18:20 crc kubenswrapper[4867]: I1212 07:18:20.359558 4867 generic.go:334] "Generic (PLEG): container finished" podID="18d7947e-f766-4149-9e15-ce48a705d4de" containerID="5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912" exitCode=0 Dec 12 07:18:20 crc kubenswrapper[4867]: I1212 07:18:20.359603 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerDied","Data":"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912"} Dec 12 07:18:21 crc kubenswrapper[4867]: I1212 07:18:21.370332 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerStarted","Data":"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945"} Dec 12 07:18:21 crc kubenswrapper[4867]: I1212 07:18:21.394415 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wd9dl" podStartSLOduration=1.697879158 podStartE2EDuration="4.394394154s" podCreationTimestamp="2025-12-12 07:18:17 +0000 UTC" firstStartedPulling="2025-12-12 07:18:18.34253255 +0000 UTC m=+1785.913913819" lastFinishedPulling="2025-12-12 07:18:21.039047546 +0000 UTC m=+1788.610428815" observedRunningTime="2025-12-12 07:18:21.388834492 +0000 UTC m=+1788.960215761" watchObservedRunningTime="2025-12-12 07:18:21.394394154 +0000 UTC m=+1788.965775423" Dec 12 07:18:24 crc kubenswrapper[4867]: I1212 07:18:24.837767 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:18:24 crc kubenswrapper[4867]: E1212 07:18:24.838358 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:18:27 crc kubenswrapper[4867]: I1212 07:18:27.395583 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:27 crc kubenswrapper[4867]: I1212 07:18:27.395923 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:27 crc kubenswrapper[4867]: I1212 07:18:27.444523 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:27 crc kubenswrapper[4867]: I1212 07:18:27.487469 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:27 crc kubenswrapper[4867]: I1212 07:18:27.678977 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:29 crc kubenswrapper[4867]: I1212 07:18:29.425361 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wd9dl" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="registry-server" containerID="cri-o://0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945" gracePeriod=2 Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.435039 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.438197 4867 generic.go:334] "Generic (PLEG): container finished" podID="18d7947e-f766-4149-9e15-ce48a705d4de" containerID="0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945" exitCode=0 Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.438255 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerDied","Data":"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945"} Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.438303 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wd9dl" event={"ID":"18d7947e-f766-4149-9e15-ce48a705d4de","Type":"ContainerDied","Data":"486ede1f6083f3d1d87835bbf08489ef49208a5ec6fd45445b4b93cd413356f0"} Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.438323 4867 scope.go:117] "RemoveContainer" containerID="0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.468034 4867 scope.go:117] "RemoveContainer" containerID="5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.489427 4867 scope.go:117] "RemoveContainer" containerID="ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.518338 4867 scope.go:117] "RemoveContainer" containerID="0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945" Dec 12 07:18:30 crc kubenswrapper[4867]: E1212 07:18:30.518781 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945\": container with ID starting with 0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945 not found: ID does not exist" containerID="0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.518814 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945"} err="failed to get container status \"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945\": rpc error: code = NotFound desc = could not find container \"0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945\": container with ID starting with 0d4efcc0c09b946418cb94f1d065b764e347255cf1cb431cee6c6567dbbe4945 not found: ID does not exist" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.518837 4867 scope.go:117] "RemoveContainer" containerID="5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912" Dec 12 07:18:30 crc kubenswrapper[4867]: E1212 07:18:30.519175 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912\": container with ID starting with 5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912 not found: ID does not exist" containerID="5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.519195 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912"} err="failed to get container status \"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912\": rpc error: code = NotFound desc = could not find container \"5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912\": container with ID starting with 5dae3544b5674c16f2043e704c30317847ffcbe6c9430065c40b6f3a8cf54912 not found: ID does not exist" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.519235 4867 scope.go:117] "RemoveContainer" containerID="ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33" Dec 12 07:18:30 crc kubenswrapper[4867]: E1212 07:18:30.519480 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33\": container with ID starting with ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33 not found: ID does not exist" containerID="ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.519508 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33"} err="failed to get container status \"ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33\": rpc error: code = NotFound desc = could not find container \"ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33\": container with ID starting with ea3b39ec15ec893da88f9a6af1dea3d1f5ed39274507c5e7756cfe906a842f33 not found: ID does not exist" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.618732 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities\") pod \"18d7947e-f766-4149-9e15-ce48a705d4de\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.618806 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqljv\" (UniqueName: \"kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv\") pod \"18d7947e-f766-4149-9e15-ce48a705d4de\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.618834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content\") pod \"18d7947e-f766-4149-9e15-ce48a705d4de\" (UID: \"18d7947e-f766-4149-9e15-ce48a705d4de\") " Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.619900 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities" (OuterVolumeSpecName: "utilities") pod "18d7947e-f766-4149-9e15-ce48a705d4de" (UID: "18d7947e-f766-4149-9e15-ce48a705d4de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.633278 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv" (OuterVolumeSpecName: "kube-api-access-jqljv") pod "18d7947e-f766-4149-9e15-ce48a705d4de" (UID: "18d7947e-f766-4149-9e15-ce48a705d4de"). InnerVolumeSpecName "kube-api-access-jqljv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.660418 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18d7947e-f766-4149-9e15-ce48a705d4de" (UID: "18d7947e-f766-4149-9e15-ce48a705d4de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.720532 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.720582 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqljv\" (UniqueName: \"kubernetes.io/projected/18d7947e-f766-4149-9e15-ce48a705d4de-kube-api-access-jqljv\") on node \"crc\" DevicePath \"\"" Dec 12 07:18:30 crc kubenswrapper[4867]: I1212 07:18:30.720598 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d7947e-f766-4149-9e15-ce48a705d4de-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:18:31 crc kubenswrapper[4867]: I1212 07:18:31.448309 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wd9dl" Dec 12 07:18:31 crc kubenswrapper[4867]: I1212 07:18:31.467477 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:31 crc kubenswrapper[4867]: I1212 07:18:31.473207 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wd9dl"] Dec 12 07:18:32 crc kubenswrapper[4867]: I1212 07:18:32.849334 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" path="/var/lib/kubelet/pods/18d7947e-f766-4149-9e15-ce48a705d4de/volumes" Dec 12 07:18:37 crc kubenswrapper[4867]: I1212 07:18:37.838293 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:18:37 crc kubenswrapper[4867]: E1212 07:18:37.838860 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:18:50 crc kubenswrapper[4867]: I1212 07:18:50.839348 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:18:50 crc kubenswrapper[4867]: E1212 07:18:50.840217 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:19:03 crc kubenswrapper[4867]: I1212 07:19:03.837683 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:19:03 crc kubenswrapper[4867]: E1212 07:19:03.838440 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:19:18 crc kubenswrapper[4867]: I1212 07:19:18.837735 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:19:18 crc kubenswrapper[4867]: E1212 07:19:18.838449 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:19:31 crc kubenswrapper[4867]: I1212 07:19:31.838495 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:19:31 crc kubenswrapper[4867]: E1212 07:19:31.839210 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:19:42 crc kubenswrapper[4867]: I1212 07:19:42.844970 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:19:42 crc kubenswrapper[4867]: E1212 07:19:42.845818 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:19:55 crc kubenswrapper[4867]: I1212 07:19:55.838487 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:19:55 crc kubenswrapper[4867]: E1212 07:19:55.839179 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:20:08 crc kubenswrapper[4867]: I1212 07:20:08.839204 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:20:08 crc kubenswrapper[4867]: E1212 07:20:08.840147 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:20:20 crc kubenswrapper[4867]: I1212 07:20:20.838114 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:20:20 crc kubenswrapper[4867]: E1212 07:20:20.838848 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:20:32 crc kubenswrapper[4867]: I1212 07:20:32.842120 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:20:32 crc kubenswrapper[4867]: E1212 07:20:32.842894 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:20:43 crc kubenswrapper[4867]: I1212 07:20:43.838279 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:20:43 crc kubenswrapper[4867]: E1212 07:20:43.839082 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:20:54 crc kubenswrapper[4867]: I1212 07:20:54.838417 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:20:54 crc kubenswrapper[4867]: E1212 07:20:54.839117 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:21:07 crc kubenswrapper[4867]: I1212 07:21:07.837719 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:21:07 crc kubenswrapper[4867]: E1212 07:21:07.838602 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:21:22 crc kubenswrapper[4867]: I1212 07:21:22.842481 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:21:22 crc kubenswrapper[4867]: E1212 07:21:22.843498 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:21:35 crc kubenswrapper[4867]: I1212 07:21:35.837902 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:21:36 crc kubenswrapper[4867]: I1212 07:21:36.785268 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72"} Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.304937 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:31 crc kubenswrapper[4867]: E1212 07:23:31.306010 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="registry-server" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.306023 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="registry-server" Dec 12 07:23:31 crc kubenswrapper[4867]: E1212 07:23:31.306052 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="extract-content" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.306059 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="extract-content" Dec 12 07:23:31 crc kubenswrapper[4867]: E1212 07:23:31.306070 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="extract-utilities" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.306077 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="extract-utilities" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.306257 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="18d7947e-f766-4149-9e15-ce48a705d4de" containerName="registry-server" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.307306 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.319821 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.430464 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8tt6\" (UniqueName: \"kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.430544 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.430581 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.531702 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.531847 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8tt6\" (UniqueName: \"kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.531885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.532446 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.532714 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.553031 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8tt6\" (UniqueName: \"kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6\") pod \"redhat-operators-qd4hr\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:31 crc kubenswrapper[4867]: I1212 07:23:31.640900 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:32 crc kubenswrapper[4867]: I1212 07:23:32.089148 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:32 crc kubenswrapper[4867]: I1212 07:23:32.826046 4867 generic.go:334] "Generic (PLEG): container finished" podID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerID="ba25a3cba4e7213392f25f6967aaeb0a33938acd6daa321795abb523625eedb3" exitCode=0 Dec 12 07:23:32 crc kubenswrapper[4867]: I1212 07:23:32.826098 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerDied","Data":"ba25a3cba4e7213392f25f6967aaeb0a33938acd6daa321795abb523625eedb3"} Dec 12 07:23:32 crc kubenswrapper[4867]: I1212 07:23:32.826437 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerStarted","Data":"d347a9f8aff9277db380ead231a468871c14d9627b6da8ad2695ad4811feb09c"} Dec 12 07:23:32 crc kubenswrapper[4867]: I1212 07:23:32.829665 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:23:33 crc kubenswrapper[4867]: I1212 07:23:33.836191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerStarted","Data":"5bc012938f6344464c777c8fdda3300e548af10978f3ed9d65a15fe0a8a23387"} Dec 12 07:23:34 crc kubenswrapper[4867]: I1212 07:23:34.846575 4867 generic.go:334] "Generic (PLEG): container finished" podID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerID="5bc012938f6344464c777c8fdda3300e548af10978f3ed9d65a15fe0a8a23387" exitCode=0 Dec 12 07:23:34 crc kubenswrapper[4867]: I1212 07:23:34.847486 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerDied","Data":"5bc012938f6344464c777c8fdda3300e548af10978f3ed9d65a15fe0a8a23387"} Dec 12 07:23:35 crc kubenswrapper[4867]: I1212 07:23:35.857090 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerStarted","Data":"4bcabae36bc1defc86a0dfd1b2a7fcd7d16a3ccb948988281d1020f63ecc880a"} Dec 12 07:23:35 crc kubenswrapper[4867]: I1212 07:23:35.875163 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qd4hr" podStartSLOduration=2.436427788 podStartE2EDuration="4.875140703s" podCreationTimestamp="2025-12-12 07:23:31 +0000 UTC" firstStartedPulling="2025-12-12 07:23:32.829377074 +0000 UTC m=+2100.400758343" lastFinishedPulling="2025-12-12 07:23:35.268089989 +0000 UTC m=+2102.839471258" observedRunningTime="2025-12-12 07:23:35.871632733 +0000 UTC m=+2103.443014002" watchObservedRunningTime="2025-12-12 07:23:35.875140703 +0000 UTC m=+2103.446521972" Dec 12 07:23:41 crc kubenswrapper[4867]: I1212 07:23:41.641767 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:41 crc kubenswrapper[4867]: I1212 07:23:41.644253 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:41 crc kubenswrapper[4867]: I1212 07:23:41.694575 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:41 crc kubenswrapper[4867]: I1212 07:23:41.933956 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:41 crc kubenswrapper[4867]: I1212 07:23:41.977612 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:43 crc kubenswrapper[4867]: I1212 07:23:43.908034 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qd4hr" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="registry-server" containerID="cri-o://4bcabae36bc1defc86a0dfd1b2a7fcd7d16a3ccb948988281d1020f63ecc880a" gracePeriod=2 Dec 12 07:23:44 crc kubenswrapper[4867]: I1212 07:23:44.924264 4867 generic.go:334] "Generic (PLEG): container finished" podID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerID="4bcabae36bc1defc86a0dfd1b2a7fcd7d16a3ccb948988281d1020f63ecc880a" exitCode=0 Dec 12 07:23:44 crc kubenswrapper[4867]: I1212 07:23:44.924349 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerDied","Data":"4bcabae36bc1defc86a0dfd1b2a7fcd7d16a3ccb948988281d1020f63ecc880a"} Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.149241 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.259712 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities\") pod \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.259873 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content\") pod \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.259923 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8tt6\" (UniqueName: \"kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6\") pod \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\" (UID: \"19ac73ea-63d7-43fb-9feb-ec3d97a826bb\") " Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.261215 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities" (OuterVolumeSpecName: "utilities") pod "19ac73ea-63d7-43fb-9feb-ec3d97a826bb" (UID: "19ac73ea-63d7-43fb-9feb-ec3d97a826bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.266281 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6" (OuterVolumeSpecName: "kube-api-access-f8tt6") pod "19ac73ea-63d7-43fb-9feb-ec3d97a826bb" (UID: "19ac73ea-63d7-43fb-9feb-ec3d97a826bb"). InnerVolumeSpecName "kube-api-access-f8tt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.361376 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8tt6\" (UniqueName: \"kubernetes.io/projected/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-kube-api-access-f8tt6\") on node \"crc\" DevicePath \"\"" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.361421 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.516036 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19ac73ea-63d7-43fb-9feb-ec3d97a826bb" (UID: "19ac73ea-63d7-43fb-9feb-ec3d97a826bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.563990 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19ac73ea-63d7-43fb-9feb-ec3d97a826bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.932902 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qd4hr" event={"ID":"19ac73ea-63d7-43fb-9feb-ec3d97a826bb","Type":"ContainerDied","Data":"d347a9f8aff9277db380ead231a468871c14d9627b6da8ad2695ad4811feb09c"} Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.932974 4867 scope.go:117] "RemoveContainer" containerID="4bcabae36bc1defc86a0dfd1b2a7fcd7d16a3ccb948988281d1020f63ecc880a" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.932998 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qd4hr" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.953160 4867 scope.go:117] "RemoveContainer" containerID="5bc012938f6344464c777c8fdda3300e548af10978f3ed9d65a15fe0a8a23387" Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.972431 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.976107 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qd4hr"] Dec 12 07:23:45 crc kubenswrapper[4867]: I1212 07:23:45.998291 4867 scope.go:117] "RemoveContainer" containerID="ba25a3cba4e7213392f25f6967aaeb0a33938acd6daa321795abb523625eedb3" Dec 12 07:23:46 crc kubenswrapper[4867]: I1212 07:23:46.846859 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" path="/var/lib/kubelet/pods/19ac73ea-63d7-43fb-9feb-ec3d97a826bb/volumes" Dec 12 07:23:58 crc kubenswrapper[4867]: I1212 07:23:58.989197 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:23:58 crc kubenswrapper[4867]: I1212 07:23:58.989746 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:24:28 crc kubenswrapper[4867]: I1212 07:24:28.989900 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:24:28 crc kubenswrapper[4867]: I1212 07:24:28.990531 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:24:58 crc kubenswrapper[4867]: I1212 07:24:58.989422 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:24:58 crc kubenswrapper[4867]: I1212 07:24:58.990119 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:24:58 crc kubenswrapper[4867]: I1212 07:24:58.990167 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:24:58 crc kubenswrapper[4867]: I1212 07:24:58.991533 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:24:58 crc kubenswrapper[4867]: I1212 07:24:58.991601 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72" gracePeriod=600 Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.492481 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72" exitCode=0 Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.492655 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72"} Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.493018 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a"} Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.493044 4867 scope.go:117] "RemoveContainer" containerID="2e39a4993e18fd96bd254adc75baa9f6b781d607c88341df5d764bea5883feda" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.794262 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:24:59 crc kubenswrapper[4867]: E1212 07:24:59.794923 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="registry-server" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.795015 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="registry-server" Dec 12 07:24:59 crc kubenswrapper[4867]: E1212 07:24:59.795109 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="extract-utilities" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.795177 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="extract-utilities" Dec 12 07:24:59 crc kubenswrapper[4867]: E1212 07:24:59.795298 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="extract-content" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.795382 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="extract-content" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.795612 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ac73ea-63d7-43fb-9feb-ec3d97a826bb" containerName="registry-server" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.796979 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.807719 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.871693 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.871755 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4bw\" (UniqueName: \"kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.871961 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.973506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.973816 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4bw\" (UniqueName: \"kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.973949 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.974058 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.974618 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:24:59 crc kubenswrapper[4867]: I1212 07:24:59.993674 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4bw\" (UniqueName: \"kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw\") pod \"community-operators-w8f7c\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:00 crc kubenswrapper[4867]: I1212 07:25:00.123792 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:00 crc kubenswrapper[4867]: I1212 07:25:00.566546 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:25:01 crc kubenswrapper[4867]: I1212 07:25:01.516192 4867 generic.go:334] "Generic (PLEG): container finished" podID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerID="8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c" exitCode=0 Dec 12 07:25:01 crc kubenswrapper[4867]: I1212 07:25:01.516347 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerDied","Data":"8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c"} Dec 12 07:25:01 crc kubenswrapper[4867]: I1212 07:25:01.516709 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerStarted","Data":"5624511fa7f0226f1deb3c9228de2947088ef48bac95c84b29b0d895fb56b79f"} Dec 12 07:25:03 crc kubenswrapper[4867]: I1212 07:25:03.534982 4867 generic.go:334] "Generic (PLEG): container finished" podID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerID="7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20" exitCode=0 Dec 12 07:25:03 crc kubenswrapper[4867]: I1212 07:25:03.535202 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerDied","Data":"7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20"} Dec 12 07:25:05 crc kubenswrapper[4867]: I1212 07:25:05.552973 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerStarted","Data":"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610"} Dec 12 07:25:05 crc kubenswrapper[4867]: I1212 07:25:05.573106 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w8f7c" podStartSLOduration=3.526709254 podStartE2EDuration="6.57308549s" podCreationTimestamp="2025-12-12 07:24:59 +0000 UTC" firstStartedPulling="2025-12-12 07:25:01.517788712 +0000 UTC m=+2189.089169981" lastFinishedPulling="2025-12-12 07:25:04.564164938 +0000 UTC m=+2192.135546217" observedRunningTime="2025-12-12 07:25:05.566747611 +0000 UTC m=+2193.138128900" watchObservedRunningTime="2025-12-12 07:25:05.57308549 +0000 UTC m=+2193.144466759" Dec 12 07:25:10 crc kubenswrapper[4867]: I1212 07:25:10.124537 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:10 crc kubenswrapper[4867]: I1212 07:25:10.124905 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:10 crc kubenswrapper[4867]: I1212 07:25:10.180695 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:10 crc kubenswrapper[4867]: I1212 07:25:10.636544 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:10 crc kubenswrapper[4867]: I1212 07:25:10.686817 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:25:12 crc kubenswrapper[4867]: I1212 07:25:12.605856 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w8f7c" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="registry-server" containerID="cri-o://d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610" gracePeriod=2 Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.493881 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.588746 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz4bw\" (UniqueName: \"kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw\") pod \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.588824 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content\") pod \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.588932 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities\") pod \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\" (UID: \"ba1b9bae-1bd8-4145-b650-cbc216d58f39\") " Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.589902 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities" (OuterVolumeSpecName: "utilities") pod "ba1b9bae-1bd8-4145-b650-cbc216d58f39" (UID: "ba1b9bae-1bd8-4145-b650-cbc216d58f39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.594084 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw" (OuterVolumeSpecName: "kube-api-access-vz4bw") pod "ba1b9bae-1bd8-4145-b650-cbc216d58f39" (UID: "ba1b9bae-1bd8-4145-b650-cbc216d58f39"). InnerVolumeSpecName "kube-api-access-vz4bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.614837 4867 generic.go:334] "Generic (PLEG): container finished" podID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerID="d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610" exitCode=0 Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.614887 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w8f7c" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.614894 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerDied","Data":"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610"} Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.615488 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w8f7c" event={"ID":"ba1b9bae-1bd8-4145-b650-cbc216d58f39","Type":"ContainerDied","Data":"5624511fa7f0226f1deb3c9228de2947088ef48bac95c84b29b0d895fb56b79f"} Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.615515 4867 scope.go:117] "RemoveContainer" containerID="d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.637541 4867 scope.go:117] "RemoveContainer" containerID="7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.651780 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba1b9bae-1bd8-4145-b650-cbc216d58f39" (UID: "ba1b9bae-1bd8-4145-b650-cbc216d58f39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.657860 4867 scope.go:117] "RemoveContainer" containerID="8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.680724 4867 scope.go:117] "RemoveContainer" containerID="d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610" Dec 12 07:25:13 crc kubenswrapper[4867]: E1212 07:25:13.681256 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610\": container with ID starting with d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610 not found: ID does not exist" containerID="d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.681289 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610"} err="failed to get container status \"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610\": rpc error: code = NotFound desc = could not find container \"d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610\": container with ID starting with d746d0071bcc970d7a266a6ad64e6f16c07768b73e227401071bbe0b635f5610 not found: ID does not exist" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.681310 4867 scope.go:117] "RemoveContainer" containerID="7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20" Dec 12 07:25:13 crc kubenswrapper[4867]: E1212 07:25:13.681926 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20\": container with ID starting with 7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20 not found: ID does not exist" containerID="7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.681970 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20"} err="failed to get container status \"7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20\": rpc error: code = NotFound desc = could not find container \"7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20\": container with ID starting with 7907ad2dea5b59101b60d05d2555d1a56d13cbb512de15bd39f70fd94cca3e20 not found: ID does not exist" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.681986 4867 scope.go:117] "RemoveContainer" containerID="8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c" Dec 12 07:25:13 crc kubenswrapper[4867]: E1212 07:25:13.682297 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c\": container with ID starting with 8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c not found: ID does not exist" containerID="8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.682325 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c"} err="failed to get container status \"8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c\": rpc error: code = NotFound desc = could not find container \"8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c\": container with ID starting with 8269db517642eacad7afe77fbd1b8a87787e5a07f1caab43f04e47c0bdbd805c not found: ID does not exist" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.690332 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.690371 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba1b9bae-1bd8-4145-b650-cbc216d58f39-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.690381 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz4bw\" (UniqueName: \"kubernetes.io/projected/ba1b9bae-1bd8-4145-b650-cbc216d58f39-kube-api-access-vz4bw\") on node \"crc\" DevicePath \"\"" Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.943036 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:25:13 crc kubenswrapper[4867]: I1212 07:25:13.948195 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w8f7c"] Dec 12 07:25:14 crc kubenswrapper[4867]: I1212 07:25:14.846079 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" path="/var/lib/kubelet/pods/ba1b9bae-1bd8-4145-b650-cbc216d58f39/volumes" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.909946 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:13 crc kubenswrapper[4867]: E1212 07:26:13.910769 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="registry-server" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.910782 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="registry-server" Dec 12 07:26:13 crc kubenswrapper[4867]: E1212 07:26:13.910796 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="extract-utilities" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.910801 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="extract-utilities" Dec 12 07:26:13 crc kubenswrapper[4867]: E1212 07:26:13.910825 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="extract-content" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.910831 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="extract-content" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.910962 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1b9bae-1bd8-4145-b650-cbc216d58f39" containerName="registry-server" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.911948 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:13 crc kubenswrapper[4867]: I1212 07:26:13.924401 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.028997 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnpdc\" (UniqueName: \"kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.029079 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.029102 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.130538 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.130582 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.130656 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnpdc\" (UniqueName: \"kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.131170 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.131310 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.151497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnpdc\" (UniqueName: \"kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc\") pod \"certified-operators-slc9k\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.241715 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:14 crc kubenswrapper[4867]: I1212 07:26:14.740457 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:15 crc kubenswrapper[4867]: I1212 07:26:15.057655 4867 generic.go:334] "Generic (PLEG): container finished" podID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerID="03a30b085a9f7fdd0194e55436395292299aac3ff705bd0f5af7fee68667807a" exitCode=0 Dec 12 07:26:15 crc kubenswrapper[4867]: I1212 07:26:15.057718 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerDied","Data":"03a30b085a9f7fdd0194e55436395292299aac3ff705bd0f5af7fee68667807a"} Dec 12 07:26:15 crc kubenswrapper[4867]: I1212 07:26:15.058028 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerStarted","Data":"66e067d256b9499b46b39e8d1329ed100691314635d2cb1b4718f5c37afa2c73"} Dec 12 07:26:16 crc kubenswrapper[4867]: I1212 07:26:16.067788 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerStarted","Data":"b8858c90700bbde91d312dfcd5db514610da2328c9869e91073df661eeaf07ed"} Dec 12 07:26:17 crc kubenswrapper[4867]: I1212 07:26:17.076085 4867 generic.go:334] "Generic (PLEG): container finished" podID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerID="b8858c90700bbde91d312dfcd5db514610da2328c9869e91073df661eeaf07ed" exitCode=0 Dec 12 07:26:17 crc kubenswrapper[4867]: I1212 07:26:17.076153 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerDied","Data":"b8858c90700bbde91d312dfcd5db514610da2328c9869e91073df661eeaf07ed"} Dec 12 07:26:18 crc kubenswrapper[4867]: I1212 07:26:18.084395 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerStarted","Data":"46c5701b16fe3d8f6d82dae55b04bdc629a683f87ff0955298c242570e887d6a"} Dec 12 07:26:18 crc kubenswrapper[4867]: I1212 07:26:18.099449 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slc9k" podStartSLOduration=2.675190536 podStartE2EDuration="5.099427263s" podCreationTimestamp="2025-12-12 07:26:13 +0000 UTC" firstStartedPulling="2025-12-12 07:26:15.059016001 +0000 UTC m=+2262.630397270" lastFinishedPulling="2025-12-12 07:26:17.483252718 +0000 UTC m=+2265.054633997" observedRunningTime="2025-12-12 07:26:18.097254752 +0000 UTC m=+2265.668636041" watchObservedRunningTime="2025-12-12 07:26:18.099427263 +0000 UTC m=+2265.670808532" Dec 12 07:26:24 crc kubenswrapper[4867]: I1212 07:26:24.242819 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:24 crc kubenswrapper[4867]: I1212 07:26:24.243552 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:24 crc kubenswrapper[4867]: I1212 07:26:24.286492 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:25 crc kubenswrapper[4867]: I1212 07:26:25.181758 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:25 crc kubenswrapper[4867]: I1212 07:26:25.222071 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:27 crc kubenswrapper[4867]: I1212 07:26:27.161549 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-slc9k" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="registry-server" containerID="cri-o://46c5701b16fe3d8f6d82dae55b04bdc629a683f87ff0955298c242570e887d6a" gracePeriod=2 Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.170223 4867 generic.go:334] "Generic (PLEG): container finished" podID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerID="46c5701b16fe3d8f6d82dae55b04bdc629a683f87ff0955298c242570e887d6a" exitCode=0 Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.170258 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerDied","Data":"46c5701b16fe3d8f6d82dae55b04bdc629a683f87ff0955298c242570e887d6a"} Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.641428 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.770736 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content\") pod \"69699c71-bca6-4fd1-b0eb-e003cb135a98\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.770889 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnpdc\" (UniqueName: \"kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc\") pod \"69699c71-bca6-4fd1-b0eb-e003cb135a98\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.770998 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities\") pod \"69699c71-bca6-4fd1-b0eb-e003cb135a98\" (UID: \"69699c71-bca6-4fd1-b0eb-e003cb135a98\") " Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.772056 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities" (OuterVolumeSpecName: "utilities") pod "69699c71-bca6-4fd1-b0eb-e003cb135a98" (UID: "69699c71-bca6-4fd1-b0eb-e003cb135a98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.777003 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc" (OuterVolumeSpecName: "kube-api-access-qnpdc") pod "69699c71-bca6-4fd1-b0eb-e003cb135a98" (UID: "69699c71-bca6-4fd1-b0eb-e003cb135a98"). InnerVolumeSpecName "kube-api-access-qnpdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.825994 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69699c71-bca6-4fd1-b0eb-e003cb135a98" (UID: "69699c71-bca6-4fd1-b0eb-e003cb135a98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.872090 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.872172 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69699c71-bca6-4fd1-b0eb-e003cb135a98-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:26:28 crc kubenswrapper[4867]: I1212 07:26:28.872188 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnpdc\" (UniqueName: \"kubernetes.io/projected/69699c71-bca6-4fd1-b0eb-e003cb135a98-kube-api-access-qnpdc\") on node \"crc\" DevicePath \"\"" Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.177690 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slc9k" event={"ID":"69699c71-bca6-4fd1-b0eb-e003cb135a98","Type":"ContainerDied","Data":"66e067d256b9499b46b39e8d1329ed100691314635d2cb1b4718f5c37afa2c73"} Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.177741 4867 scope.go:117] "RemoveContainer" containerID="46c5701b16fe3d8f6d82dae55b04bdc629a683f87ff0955298c242570e887d6a" Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.177741 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slc9k" Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.198129 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.204045 4867 scope.go:117] "RemoveContainer" containerID="b8858c90700bbde91d312dfcd5db514610da2328c9869e91073df661eeaf07ed" Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.205700 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-slc9k"] Dec 12 07:26:29 crc kubenswrapper[4867]: I1212 07:26:29.223894 4867 scope.go:117] "RemoveContainer" containerID="03a30b085a9f7fdd0194e55436395292299aac3ff705bd0f5af7fee68667807a" Dec 12 07:26:30 crc kubenswrapper[4867]: I1212 07:26:30.845602 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" path="/var/lib/kubelet/pods/69699c71-bca6-4fd1-b0eb-e003cb135a98/volumes" Dec 12 07:27:28 crc kubenswrapper[4867]: I1212 07:27:28.989449 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:27:28 crc kubenswrapper[4867]: I1212 07:27:28.990078 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:27:58 crc kubenswrapper[4867]: I1212 07:27:58.988832 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:27:58 crc kubenswrapper[4867]: I1212 07:27:58.989436 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:28:28 crc kubenswrapper[4867]: I1212 07:28:28.988690 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:28:28 crc kubenswrapper[4867]: I1212 07:28:28.989274 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:28:28 crc kubenswrapper[4867]: I1212 07:28:28.989339 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:28:28 crc kubenswrapper[4867]: I1212 07:28:28.989861 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:28:28 crc kubenswrapper[4867]: I1212 07:28:28.989905 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" gracePeriod=600 Dec 12 07:28:29 crc kubenswrapper[4867]: E1212 07:28:29.111721 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:28:29 crc kubenswrapper[4867]: I1212 07:28:29.982758 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" exitCode=0 Dec 12 07:28:29 crc kubenswrapper[4867]: I1212 07:28:29.982873 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a"} Dec 12 07:28:29 crc kubenswrapper[4867]: I1212 07:28:29.982942 4867 scope.go:117] "RemoveContainer" containerID="e973100f4e0edadfb31b4c1b13e91937078caadac8dc55a328e48d1b8ba8ee72" Dec 12 07:28:29 crc kubenswrapper[4867]: I1212 07:28:29.983456 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:28:29 crc kubenswrapper[4867]: E1212 07:28:29.983689 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:28:41 crc kubenswrapper[4867]: I1212 07:28:41.838153 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:28:41 crc kubenswrapper[4867]: E1212 07:28:41.839857 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:28:52 crc kubenswrapper[4867]: I1212 07:28:52.843888 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:28:52 crc kubenswrapper[4867]: E1212 07:28:52.844739 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.129583 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:28:56 crc kubenswrapper[4867]: E1212 07:28:56.130409 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="extract-content" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.130422 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="extract-content" Dec 12 07:28:56 crc kubenswrapper[4867]: E1212 07:28:56.130441 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="registry-server" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.130447 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="registry-server" Dec 12 07:28:56 crc kubenswrapper[4867]: E1212 07:28:56.130459 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="extract-utilities" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.130465 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="extract-utilities" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.130627 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="69699c71-bca6-4fd1-b0eb-e003cb135a98" containerName="registry-server" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.131606 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.176826 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.270192 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvknw\" (UniqueName: \"kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.270259 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.270529 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.372012 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvknw\" (UniqueName: \"kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.372070 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.372132 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.372804 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.372996 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.395739 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvknw\" (UniqueName: \"kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw\") pod \"redhat-marketplace-lscwd\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.470084 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:28:56 crc kubenswrapper[4867]: I1212 07:28:56.946144 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:28:57 crc kubenswrapper[4867]: I1212 07:28:57.185959 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerID="4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96" exitCode=0 Dec 12 07:28:57 crc kubenswrapper[4867]: I1212 07:28:57.186085 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerDied","Data":"4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96"} Dec 12 07:28:57 crc kubenswrapper[4867]: I1212 07:28:57.186543 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerStarted","Data":"971f2e18e226b6bf2c5c5af3e50b4ad5bd43129509b5158c4643853262276c66"} Dec 12 07:28:57 crc kubenswrapper[4867]: I1212 07:28:57.188455 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:28:58 crc kubenswrapper[4867]: I1212 07:28:58.194999 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerID="26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d" exitCode=0 Dec 12 07:28:58 crc kubenswrapper[4867]: I1212 07:28:58.195280 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerDied","Data":"26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d"} Dec 12 07:28:59 crc kubenswrapper[4867]: I1212 07:28:59.203060 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerStarted","Data":"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c"} Dec 12 07:28:59 crc kubenswrapper[4867]: I1212 07:28:59.229027 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lscwd" podStartSLOduration=1.749492638 podStartE2EDuration="3.229006939s" podCreationTimestamp="2025-12-12 07:28:56 +0000 UTC" firstStartedPulling="2025-12-12 07:28:57.188169435 +0000 UTC m=+2424.759550704" lastFinishedPulling="2025-12-12 07:28:58.667683746 +0000 UTC m=+2426.239065005" observedRunningTime="2025-12-12 07:28:59.22397786 +0000 UTC m=+2426.795359129" watchObservedRunningTime="2025-12-12 07:28:59.229006939 +0000 UTC m=+2426.800388208" Dec 12 07:29:06 crc kubenswrapper[4867]: I1212 07:29:06.470557 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:06 crc kubenswrapper[4867]: I1212 07:29:06.471198 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:06 crc kubenswrapper[4867]: I1212 07:29:06.517898 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:07 crc kubenswrapper[4867]: I1212 07:29:07.319821 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:07 crc kubenswrapper[4867]: I1212 07:29:07.370525 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:29:07 crc kubenswrapper[4867]: I1212 07:29:07.837691 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:29:07 crc kubenswrapper[4867]: E1212 07:29:07.837943 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:29:09 crc kubenswrapper[4867]: I1212 07:29:09.290593 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lscwd" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="registry-server" containerID="cri-o://0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c" gracePeriod=2 Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.182037 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.357158 4867 generic.go:334] "Generic (PLEG): container finished" podID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerID="0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c" exitCode=0 Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.357215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerDied","Data":"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c"} Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.357266 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lscwd" event={"ID":"8a38f1c7-2301-4ad4-9fe1-46be1c71c352","Type":"ContainerDied","Data":"971f2e18e226b6bf2c5c5af3e50b4ad5bd43129509b5158c4643853262276c66"} Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.357299 4867 scope.go:117] "RemoveContainer" containerID="0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.357515 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lscwd" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.379465 4867 scope.go:117] "RemoveContainer" containerID="26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.398588 4867 scope.go:117] "RemoveContainer" containerID="4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.432079 4867 scope.go:117] "RemoveContainer" containerID="0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c" Dec 12 07:29:10 crc kubenswrapper[4867]: E1212 07:29:10.434505 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c\": container with ID starting with 0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c not found: ID does not exist" containerID="0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.434559 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c"} err="failed to get container status \"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c\": rpc error: code = NotFound desc = could not find container \"0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c\": container with ID starting with 0d83ee18f70f548299b5bbbe48a7042b8fd8ce0d7d5ef7272579e7bd190ec91c not found: ID does not exist" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.434590 4867 scope.go:117] "RemoveContainer" containerID="26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d" Dec 12 07:29:10 crc kubenswrapper[4867]: E1212 07:29:10.438795 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d\": container with ID starting with 26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d not found: ID does not exist" containerID="26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.438873 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d"} err="failed to get container status \"26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d\": rpc error: code = NotFound desc = could not find container \"26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d\": container with ID starting with 26cd2be3e49463116cbc256111cb879477ab7b8ef056125fe1661c5e20ae980d not found: ID does not exist" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.438919 4867 scope.go:117] "RemoveContainer" containerID="4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96" Dec 12 07:29:10 crc kubenswrapper[4867]: E1212 07:29:10.439583 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96\": container with ID starting with 4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96 not found: ID does not exist" containerID="4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.439618 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96"} err="failed to get container status \"4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96\": rpc error: code = NotFound desc = could not find container \"4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96\": container with ID starting with 4fbc8b5a9fbff323416d1e6f84c6c23d1d2611d82483916f8010c7d50175de96 not found: ID does not exist" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.443419 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvknw\" (UniqueName: \"kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw\") pod \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.443797 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities\") pod \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.443895 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content\") pod \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\" (UID: \"8a38f1c7-2301-4ad4-9fe1-46be1c71c352\") " Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.444731 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities" (OuterVolumeSpecName: "utilities") pod "8a38f1c7-2301-4ad4-9fe1-46be1c71c352" (UID: "8a38f1c7-2301-4ad4-9fe1-46be1c71c352"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.451129 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw" (OuterVolumeSpecName: "kube-api-access-kvknw") pod "8a38f1c7-2301-4ad4-9fe1-46be1c71c352" (UID: "8a38f1c7-2301-4ad4-9fe1-46be1c71c352"). InnerVolumeSpecName "kube-api-access-kvknw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.464922 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a38f1c7-2301-4ad4-9fe1-46be1c71c352" (UID: "8a38f1c7-2301-4ad4-9fe1-46be1c71c352"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.545579 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.545612 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.545624 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvknw\" (UniqueName: \"kubernetes.io/projected/8a38f1c7-2301-4ad4-9fe1-46be1c71c352-kube-api-access-kvknw\") on node \"crc\" DevicePath \"\"" Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.687128 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.694523 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lscwd"] Dec 12 07:29:10 crc kubenswrapper[4867]: I1212 07:29:10.847814 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" path="/var/lib/kubelet/pods/8a38f1c7-2301-4ad4-9fe1-46be1c71c352/volumes" Dec 12 07:29:19 crc kubenswrapper[4867]: I1212 07:29:19.838076 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:29:19 crc kubenswrapper[4867]: E1212 07:29:19.838789 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:29:34 crc kubenswrapper[4867]: I1212 07:29:34.838709 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:29:34 crc kubenswrapper[4867]: E1212 07:29:34.840769 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:29:48 crc kubenswrapper[4867]: I1212 07:29:48.838845 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:29:48 crc kubenswrapper[4867]: E1212 07:29:48.839628 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.140267 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc"] Dec 12 07:30:00 crc kubenswrapper[4867]: E1212 07:30:00.141137 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="extract-content" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.141151 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="extract-content" Dec 12 07:30:00 crc kubenswrapper[4867]: E1212 07:30:00.141164 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="extract-utilities" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.141172 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="extract-utilities" Dec 12 07:30:00 crc kubenswrapper[4867]: E1212 07:30:00.141195 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="registry-server" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.141202 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="registry-server" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.141415 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a38f1c7-2301-4ad4-9fe1-46be1c71c352" containerName="registry-server" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.142060 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.144265 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.145210 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.150427 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc"] Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.203592 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.203683 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fspjz\" (UniqueName: \"kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.203725 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.304506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.304565 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fspjz\" (UniqueName: \"kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.304601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.305718 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.310892 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.321393 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fspjz\" (UniqueName: \"kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz\") pod \"collect-profiles-29425410-tpdxc\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.466813 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:00 crc kubenswrapper[4867]: I1212 07:30:00.889043 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc"] Dec 12 07:30:01 crc kubenswrapper[4867]: I1212 07:30:01.728873 4867 generic.go:334] "Generic (PLEG): container finished" podID="b79b364c-f390-4159-a7a4-25f976acdd87" containerID="5ab635b37604298b1c5cfaee2df09d922f07070c79a48401bc5f0151a08a58dd" exitCode=0 Dec 12 07:30:01 crc kubenswrapper[4867]: I1212 07:30:01.728979 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" event={"ID":"b79b364c-f390-4159-a7a4-25f976acdd87","Type":"ContainerDied","Data":"5ab635b37604298b1c5cfaee2df09d922f07070c79a48401bc5f0151a08a58dd"} Dec 12 07:30:01 crc kubenswrapper[4867]: I1212 07:30:01.729169 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" event={"ID":"b79b364c-f390-4159-a7a4-25f976acdd87","Type":"ContainerStarted","Data":"caac4d154af53c5de6a395012921fe2b34d11b2ca07103ccade0b79d4f0403f5"} Dec 12 07:30:01 crc kubenswrapper[4867]: I1212 07:30:01.838489 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:30:01 crc kubenswrapper[4867]: E1212 07:30:01.838893 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:30:02 crc kubenswrapper[4867]: I1212 07:30:02.990098 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.154807 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fspjz\" (UniqueName: \"kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz\") pod \"b79b364c-f390-4159-a7a4-25f976acdd87\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.154956 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume\") pod \"b79b364c-f390-4159-a7a4-25f976acdd87\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.155000 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume\") pod \"b79b364c-f390-4159-a7a4-25f976acdd87\" (UID: \"b79b364c-f390-4159-a7a4-25f976acdd87\") " Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.156124 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume" (OuterVolumeSpecName: "config-volume") pod "b79b364c-f390-4159-a7a4-25f976acdd87" (UID: "b79b364c-f390-4159-a7a4-25f976acdd87"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.161012 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz" (OuterVolumeSpecName: "kube-api-access-fspjz") pod "b79b364c-f390-4159-a7a4-25f976acdd87" (UID: "b79b364c-f390-4159-a7a4-25f976acdd87"). InnerVolumeSpecName "kube-api-access-fspjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.162426 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b79b364c-f390-4159-a7a4-25f976acdd87" (UID: "b79b364c-f390-4159-a7a4-25f976acdd87"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.256910 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b79b364c-f390-4159-a7a4-25f976acdd87-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.256946 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b79b364c-f390-4159-a7a4-25f976acdd87-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.256956 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fspjz\" (UniqueName: \"kubernetes.io/projected/b79b364c-f390-4159-a7a4-25f976acdd87-kube-api-access-fspjz\") on node \"crc\" DevicePath \"\"" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.746033 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" event={"ID":"b79b364c-f390-4159-a7a4-25f976acdd87","Type":"ContainerDied","Data":"caac4d154af53c5de6a395012921fe2b34d11b2ca07103ccade0b79d4f0403f5"} Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.746086 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caac4d154af53c5de6a395012921fe2b34d11b2ca07103ccade0b79d4f0403f5" Dec 12 07:30:03 crc kubenswrapper[4867]: I1212 07:30:03.746117 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc" Dec 12 07:30:04 crc kubenswrapper[4867]: I1212 07:30:04.059976 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w"] Dec 12 07:30:04 crc kubenswrapper[4867]: I1212 07:30:04.067601 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425365-5s62w"] Dec 12 07:30:04 crc kubenswrapper[4867]: I1212 07:30:04.852295 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f8fec8-9201-4083-8948-6c2958ef5ec7" path="/var/lib/kubelet/pods/f8f8fec8-9201-4083-8948-6c2958ef5ec7/volumes" Dec 12 07:30:15 crc kubenswrapper[4867]: I1212 07:30:15.837888 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:30:15 crc kubenswrapper[4867]: E1212 07:30:15.838678 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:30:27 crc kubenswrapper[4867]: I1212 07:30:27.838661 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:30:27 crc kubenswrapper[4867]: E1212 07:30:27.839448 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:30:38 crc kubenswrapper[4867]: I1212 07:30:38.838445 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:30:38 crc kubenswrapper[4867]: E1212 07:30:38.839062 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:30:42 crc kubenswrapper[4867]: I1212 07:30:42.809948 4867 scope.go:117] "RemoveContainer" containerID="9b56c8d56d3ed1c611eb7adf9794fc00705e4ecb54e06bc9c8b61927059968ef" Dec 12 07:30:49 crc kubenswrapper[4867]: I1212 07:30:49.839264 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:30:49 crc kubenswrapper[4867]: E1212 07:30:49.840621 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:31:02 crc kubenswrapper[4867]: I1212 07:31:02.841484 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:31:02 crc kubenswrapper[4867]: E1212 07:31:02.842326 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:31:14 crc kubenswrapper[4867]: I1212 07:31:14.838359 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:31:14 crc kubenswrapper[4867]: E1212 07:31:14.839647 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:31:28 crc kubenswrapper[4867]: I1212 07:31:28.838634 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:31:28 crc kubenswrapper[4867]: E1212 07:31:28.839331 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:31:42 crc kubenswrapper[4867]: I1212 07:31:42.843019 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:31:42 crc kubenswrapper[4867]: E1212 07:31:42.843844 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:31:55 crc kubenswrapper[4867]: I1212 07:31:55.838737 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:31:55 crc kubenswrapper[4867]: E1212 07:31:55.851782 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:32:08 crc kubenswrapper[4867]: I1212 07:32:08.838032 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:32:08 crc kubenswrapper[4867]: E1212 07:32:08.840163 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:32:20 crc kubenswrapper[4867]: I1212 07:32:20.838360 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:32:20 crc kubenswrapper[4867]: E1212 07:32:20.839571 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:32:32 crc kubenswrapper[4867]: I1212 07:32:32.837954 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:32:32 crc kubenswrapper[4867]: E1212 07:32:32.838679 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:32:45 crc kubenswrapper[4867]: I1212 07:32:45.838821 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:32:45 crc kubenswrapper[4867]: E1212 07:32:45.839698 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:32:56 crc kubenswrapper[4867]: I1212 07:32:56.837621 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:32:56 crc kubenswrapper[4867]: E1212 07:32:56.838198 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:33:09 crc kubenswrapper[4867]: I1212 07:33:09.838071 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:33:09 crc kubenswrapper[4867]: E1212 07:33:09.838921 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:33:20 crc kubenswrapper[4867]: I1212 07:33:20.837880 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:33:20 crc kubenswrapper[4867]: E1212 07:33:20.838566 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:33:31 crc kubenswrapper[4867]: I1212 07:33:31.838024 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.392682 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a"} Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.615569 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:32 crc kubenswrapper[4867]: E1212 07:33:32.618267 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79b364c-f390-4159-a7a4-25f976acdd87" containerName="collect-profiles" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.618975 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79b364c-f390-4159-a7a4-25f976acdd87" containerName="collect-profiles" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.619687 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79b364c-f390-4159-a7a4-25f976acdd87" containerName="collect-profiles" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.621848 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.630619 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.761383 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.761626 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8wnt\" (UniqueName: \"kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.761679 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.864086 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.864164 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8wnt\" (UniqueName: \"kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.864193 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.864893 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.865025 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.896084 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8wnt\" (UniqueName: \"kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt\") pod \"redhat-operators-bf24v\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:32 crc kubenswrapper[4867]: I1212 07:33:32.994907 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:33 crc kubenswrapper[4867]: I1212 07:33:33.429102 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:34 crc kubenswrapper[4867]: I1212 07:33:34.409103 4867 generic.go:334] "Generic (PLEG): container finished" podID="54a76975-adf2-4eac-9408-5356ebdd7151" containerID="6abbc47980744a2c32fa266d2db90b8bdb55fccbcffb74d156bcc55b6dc70e11" exitCode=0 Dec 12 07:33:34 crc kubenswrapper[4867]: I1212 07:33:34.409155 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerDied","Data":"6abbc47980744a2c32fa266d2db90b8bdb55fccbcffb74d156bcc55b6dc70e11"} Dec 12 07:33:34 crc kubenswrapper[4867]: I1212 07:33:34.409684 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerStarted","Data":"2e3a0c7ef7d1e0d8b3f900e9ca9172b0e472bda1d5180a777eec510e5710b892"} Dec 12 07:33:35 crc kubenswrapper[4867]: I1212 07:33:35.421125 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerStarted","Data":"b8b3860b4966becb91a94a830b31ba175e1c5117f5d2927c5b81fbf4198d455d"} Dec 12 07:33:36 crc kubenswrapper[4867]: I1212 07:33:36.431457 4867 generic.go:334] "Generic (PLEG): container finished" podID="54a76975-adf2-4eac-9408-5356ebdd7151" containerID="b8b3860b4966becb91a94a830b31ba175e1c5117f5d2927c5b81fbf4198d455d" exitCode=0 Dec 12 07:33:36 crc kubenswrapper[4867]: I1212 07:33:36.431529 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerDied","Data":"b8b3860b4966becb91a94a830b31ba175e1c5117f5d2927c5b81fbf4198d455d"} Dec 12 07:33:37 crc kubenswrapper[4867]: I1212 07:33:37.441998 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerStarted","Data":"544b159b72d6392eb8c32b0713458666a22d3f05a4105ad4208696310bd3bc47"} Dec 12 07:33:37 crc kubenswrapper[4867]: I1212 07:33:37.460499 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bf24v" podStartSLOduration=2.949434696 podStartE2EDuration="5.460481382s" podCreationTimestamp="2025-12-12 07:33:32 +0000 UTC" firstStartedPulling="2025-12-12 07:33:34.410715883 +0000 UTC m=+2701.982097152" lastFinishedPulling="2025-12-12 07:33:36.921762569 +0000 UTC m=+2704.493143838" observedRunningTime="2025-12-12 07:33:37.457691833 +0000 UTC m=+2705.029073102" watchObservedRunningTime="2025-12-12 07:33:37.460481382 +0000 UTC m=+2705.031862651" Dec 12 07:33:42 crc kubenswrapper[4867]: I1212 07:33:42.996081 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:42 crc kubenswrapper[4867]: I1212 07:33:42.996677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:43 crc kubenswrapper[4867]: I1212 07:33:43.035058 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:43 crc kubenswrapper[4867]: I1212 07:33:43.525421 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:43 crc kubenswrapper[4867]: I1212 07:33:43.568981 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:45 crc kubenswrapper[4867]: I1212 07:33:45.504669 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bf24v" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="registry-server" containerID="cri-o://544b159b72d6392eb8c32b0713458666a22d3f05a4105ad4208696310bd3bc47" gracePeriod=2 Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.549905 4867 generic.go:334] "Generic (PLEG): container finished" podID="54a76975-adf2-4eac-9408-5356ebdd7151" containerID="544b159b72d6392eb8c32b0713458666a22d3f05a4105ad4208696310bd3bc47" exitCode=0 Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.550008 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerDied","Data":"544b159b72d6392eb8c32b0713458666a22d3f05a4105ad4208696310bd3bc47"} Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.649024 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.796690 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content\") pod \"54a76975-adf2-4eac-9408-5356ebdd7151\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.796757 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8wnt\" (UniqueName: \"kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt\") pod \"54a76975-adf2-4eac-9408-5356ebdd7151\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.796865 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities\") pod \"54a76975-adf2-4eac-9408-5356ebdd7151\" (UID: \"54a76975-adf2-4eac-9408-5356ebdd7151\") " Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.797910 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities" (OuterVolumeSpecName: "utilities") pod "54a76975-adf2-4eac-9408-5356ebdd7151" (UID: "54a76975-adf2-4eac-9408-5356ebdd7151"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.803099 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt" (OuterVolumeSpecName: "kube-api-access-j8wnt") pod "54a76975-adf2-4eac-9408-5356ebdd7151" (UID: "54a76975-adf2-4eac-9408-5356ebdd7151"). InnerVolumeSpecName "kube-api-access-j8wnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.898705 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.898752 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8wnt\" (UniqueName: \"kubernetes.io/projected/54a76975-adf2-4eac-9408-5356ebdd7151-kube-api-access-j8wnt\") on node \"crc\" DevicePath \"\"" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.919852 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54a76975-adf2-4eac-9408-5356ebdd7151" (UID: "54a76975-adf2-4eac-9408-5356ebdd7151"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:33:49 crc kubenswrapper[4867]: I1212 07:33:49.999850 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54a76975-adf2-4eac-9408-5356ebdd7151-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.560486 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf24v" event={"ID":"54a76975-adf2-4eac-9408-5356ebdd7151","Type":"ContainerDied","Data":"2e3a0c7ef7d1e0d8b3f900e9ca9172b0e472bda1d5180a777eec510e5710b892"} Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.560610 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf24v" Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.560814 4867 scope.go:117] "RemoveContainer" containerID="544b159b72d6392eb8c32b0713458666a22d3f05a4105ad4208696310bd3bc47" Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.583922 4867 scope.go:117] "RemoveContainer" containerID="b8b3860b4966becb91a94a830b31ba175e1c5117f5d2927c5b81fbf4198d455d" Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.591043 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.596491 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bf24v"] Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.629147 4867 scope.go:117] "RemoveContainer" containerID="6abbc47980744a2c32fa266d2db90b8bdb55fccbcffb74d156bcc55b6dc70e11" Dec 12 07:33:50 crc kubenswrapper[4867]: I1212 07:33:50.848542 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" path="/var/lib/kubelet/pods/54a76975-adf2-4eac-9408-5356ebdd7151/volumes" Dec 12 07:35:58 crc kubenswrapper[4867]: I1212 07:35:58.989130 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:35:58 crc kubenswrapper[4867]: I1212 07:35:58.989935 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:36:28 crc kubenswrapper[4867]: I1212 07:36:28.988892 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:36:28 crc kubenswrapper[4867]: I1212 07:36:28.989537 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:36:58 crc kubenswrapper[4867]: I1212 07:36:58.989332 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:36:58 crc kubenswrapper[4867]: I1212 07:36:58.990031 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:36:58 crc kubenswrapper[4867]: I1212 07:36:58.990097 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:36:58 crc kubenswrapper[4867]: I1212 07:36:58.991082 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:36:58 crc kubenswrapper[4867]: I1212 07:36:58.991173 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a" gracePeriod=600 Dec 12 07:36:59 crc kubenswrapper[4867]: I1212 07:36:59.927562 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a" exitCode=0 Dec 12 07:36:59 crc kubenswrapper[4867]: I1212 07:36:59.927638 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a"} Dec 12 07:36:59 crc kubenswrapper[4867]: I1212 07:36:59.928121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b"} Dec 12 07:36:59 crc kubenswrapper[4867]: I1212 07:36:59.928147 4867 scope.go:117] "RemoveContainer" containerID="9ed53c5400747e8c6e7c4afce6e9fda36efe0c7aed3dd4c9f5c462b0b5102c8a" Dec 12 07:39:28 crc kubenswrapper[4867]: I1212 07:39:28.989341 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:39:28 crc kubenswrapper[4867]: I1212 07:39:28.989833 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:39:58 crc kubenswrapper[4867]: I1212 07:39:58.989464 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:39:58 crc kubenswrapper[4867]: I1212 07:39:58.989968 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:40:28 crc kubenswrapper[4867]: I1212 07:40:28.988580 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:40:28 crc kubenswrapper[4867]: I1212 07:40:28.989151 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:40:28 crc kubenswrapper[4867]: I1212 07:40:28.989194 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:40:28 crc kubenswrapper[4867]: I1212 07:40:28.989867 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:40:28 crc kubenswrapper[4867]: I1212 07:40:28.989971 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" gracePeriod=600 Dec 12 07:40:29 crc kubenswrapper[4867]: E1212 07:40:29.113390 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:40:29 crc kubenswrapper[4867]: I1212 07:40:29.377504 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" exitCode=0 Dec 12 07:40:29 crc kubenswrapper[4867]: I1212 07:40:29.377551 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b"} Dec 12 07:40:29 crc kubenswrapper[4867]: I1212 07:40:29.377586 4867 scope.go:117] "RemoveContainer" containerID="eac738ed634dbf41b88471fded5e7dbda6ea51e398b341b77b0e4de393897c8a" Dec 12 07:40:29 crc kubenswrapper[4867]: I1212 07:40:29.378493 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:40:29 crc kubenswrapper[4867]: E1212 07:40:29.378956 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:40:41 crc kubenswrapper[4867]: I1212 07:40:41.838958 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:40:41 crc kubenswrapper[4867]: E1212 07:40:41.840495 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:40:52 crc kubenswrapper[4867]: I1212 07:40:52.842652 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:40:52 crc kubenswrapper[4867]: E1212 07:40:52.844096 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:41:05 crc kubenswrapper[4867]: I1212 07:41:05.838096 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:41:05 crc kubenswrapper[4867]: E1212 07:41:05.839463 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:41:18 crc kubenswrapper[4867]: I1212 07:41:18.838036 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:41:18 crc kubenswrapper[4867]: E1212 07:41:18.838664 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:41:33 crc kubenswrapper[4867]: I1212 07:41:33.838307 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:41:33 crc kubenswrapper[4867]: E1212 07:41:33.839011 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:41:45 crc kubenswrapper[4867]: I1212 07:41:45.838622 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:41:45 crc kubenswrapper[4867]: E1212 07:41:45.839488 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:41:58 crc kubenswrapper[4867]: I1212 07:41:58.837870 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:41:58 crc kubenswrapper[4867]: E1212 07:41:58.838773 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:42:09 crc kubenswrapper[4867]: I1212 07:42:09.838177 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:42:09 crc kubenswrapper[4867]: E1212 07:42:09.838965 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:42:23 crc kubenswrapper[4867]: I1212 07:42:23.839110 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:42:23 crc kubenswrapper[4867]: E1212 07:42:23.839910 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:42:36 crc kubenswrapper[4867]: I1212 07:42:36.838654 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:42:36 crc kubenswrapper[4867]: E1212 07:42:36.839418 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.833764 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:42:49 crc kubenswrapper[4867]: E1212 07:42:49.834655 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="registry-server" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.834670 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="registry-server" Dec 12 07:42:49 crc kubenswrapper[4867]: E1212 07:42:49.834699 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="extract-content" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.834706 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="extract-content" Dec 12 07:42:49 crc kubenswrapper[4867]: E1212 07:42:49.834718 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="extract-utilities" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.834724 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="extract-utilities" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.834855 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a76975-adf2-4eac-9408-5356ebdd7151" containerName="registry-server" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.837719 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:49 crc kubenswrapper[4867]: I1212 07:42:49.851115 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.030825 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.032303 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.035687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26rz9\" (UniqueName: \"kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.036216 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.037008 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.074272 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.141649 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26rz9\" (UniqueName: \"kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.142048 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.142109 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.142145 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.142168 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.142201 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdvt2\" (UniqueName: \"kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.143087 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.143409 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.164382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26rz9\" (UniqueName: \"kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9\") pod \"community-operators-2f9h2\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.243810 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.243909 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.243958 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdvt2\" (UniqueName: \"kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.244485 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.244514 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.262672 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdvt2\" (UniqueName: \"kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2\") pod \"certified-operators-m5pv5\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.372271 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.457805 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.765016 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:42:50 crc kubenswrapper[4867]: I1212 07:42:50.870934 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:42:50 crc kubenswrapper[4867]: W1212 07:42:50.880993 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod496cb354_7260_4c20_8e90_6a0ff6ba944c.slice/crio-40176b5af1795b6cb09d3fd5f584187027b90d3359dcbc0ab8141d1123c1eccd WatchSource:0}: Error finding container 40176b5af1795b6cb09d3fd5f584187027b90d3359dcbc0ab8141d1123c1eccd: Status 404 returned error can't find the container with id 40176b5af1795b6cb09d3fd5f584187027b90d3359dcbc0ab8141d1123c1eccd Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.331351 4867 generic.go:334] "Generic (PLEG): container finished" podID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerID="2ecb2ac2b0aa08f42f93e719213b99365a0ae6445c9fe67fdc08e3df9796ce1d" exitCode=0 Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.331437 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerDied","Data":"2ecb2ac2b0aa08f42f93e719213b99365a0ae6445c9fe67fdc08e3df9796ce1d"} Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.331777 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerStarted","Data":"5859e142b55ee43c5a44a7d4b87012c75572658d91d156beebb8d70f07342579"} Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.333303 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.333683 4867 generic.go:334] "Generic (PLEG): container finished" podID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerID="4f279079f7be1f15dc974fb0e77e5a39de6228a97757b9f32844d543887955c4" exitCode=0 Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.333720 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerDied","Data":"4f279079f7be1f15dc974fb0e77e5a39de6228a97757b9f32844d543887955c4"} Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.333744 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerStarted","Data":"40176b5af1795b6cb09d3fd5f584187027b90d3359dcbc0ab8141d1123c1eccd"} Dec 12 07:42:51 crc kubenswrapper[4867]: I1212 07:42:51.838147 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:42:51 crc kubenswrapper[4867]: E1212 07:42:51.838471 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.241336 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.250002 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.257351 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.376657 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.376804 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq65c\" (UniqueName: \"kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.376894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.478375 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq65c\" (UniqueName: \"kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.478443 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.478492 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.478994 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.479513 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.497251 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq65c\" (UniqueName: \"kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c\") pod \"redhat-marketplace-l2kjg\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.609094 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:42:52 crc kubenswrapper[4867]: I1212 07:42:52.828719 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.354805 4867 generic.go:334] "Generic (PLEG): container finished" podID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerID="c0759090b69cee5eb37c90bcc1c2ca426126efbbc68c5c8157716687b86685ea" exitCode=0 Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.354873 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerDied","Data":"c0759090b69cee5eb37c90bcc1c2ca426126efbbc68c5c8157716687b86685ea"} Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.358299 4867 generic.go:334] "Generic (PLEG): container finished" podID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerID="77c07622cb43f89f8b3ba9ec5997f9fd2b8eab6d645641a570e01bb02f3b9146" exitCode=0 Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.358353 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerDied","Data":"77c07622cb43f89f8b3ba9ec5997f9fd2b8eab6d645641a570e01bb02f3b9146"} Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.359884 4867 generic.go:334] "Generic (PLEG): container finished" podID="66f01148-89f0-4473-96e4-ca3050892aab" containerID="0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf" exitCode=0 Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.359906 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerDied","Data":"0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf"} Dec 12 07:42:53 crc kubenswrapper[4867]: I1212 07:42:53.359922 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerStarted","Data":"5a1381637c1030ba53e5ee68d19d650f8a236d5e38ef89cfac1b4d9e146042b6"} Dec 12 07:42:54 crc kubenswrapper[4867]: I1212 07:42:54.369523 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerStarted","Data":"c88a697cad0acfe56b8081e914ee12740f007aabe67852d17711460ca4c28a75"} Dec 12 07:42:54 crc kubenswrapper[4867]: I1212 07:42:54.374185 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerStarted","Data":"5f16d02d6b2458a4101c7fe14b4488f271370c717e7fdb097938cad6d6e907d4"} Dec 12 07:42:54 crc kubenswrapper[4867]: I1212 07:42:54.379154 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerStarted","Data":"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57"} Dec 12 07:42:54 crc kubenswrapper[4867]: I1212 07:42:54.393095 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2f9h2" podStartSLOduration=2.977333532 podStartE2EDuration="5.393076128s" podCreationTimestamp="2025-12-12 07:42:49 +0000 UTC" firstStartedPulling="2025-12-12 07:42:51.332694017 +0000 UTC m=+3258.904075286" lastFinishedPulling="2025-12-12 07:42:53.748436613 +0000 UTC m=+3261.319817882" observedRunningTime="2025-12-12 07:42:54.388472974 +0000 UTC m=+3261.959854243" watchObservedRunningTime="2025-12-12 07:42:54.393076128 +0000 UTC m=+3261.964457397" Dec 12 07:42:54 crc kubenswrapper[4867]: I1212 07:42:54.445100 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5pv5" podStartSLOduration=2.019905463 podStartE2EDuration="4.445067099s" podCreationTimestamp="2025-12-12 07:42:50 +0000 UTC" firstStartedPulling="2025-12-12 07:42:51.334986874 +0000 UTC m=+3258.906368143" lastFinishedPulling="2025-12-12 07:42:53.76014851 +0000 UTC m=+3261.331529779" observedRunningTime="2025-12-12 07:42:54.436802087 +0000 UTC m=+3262.008183356" watchObservedRunningTime="2025-12-12 07:42:54.445067099 +0000 UTC m=+3262.016448368" Dec 12 07:42:55 crc kubenswrapper[4867]: I1212 07:42:55.387286 4867 generic.go:334] "Generic (PLEG): container finished" podID="66f01148-89f0-4473-96e4-ca3050892aab" containerID="348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57" exitCode=0 Dec 12 07:42:55 crc kubenswrapper[4867]: I1212 07:42:55.387386 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerDied","Data":"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57"} Dec 12 07:42:56 crc kubenswrapper[4867]: I1212 07:42:56.401856 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerStarted","Data":"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2"} Dec 12 07:42:56 crc kubenswrapper[4867]: I1212 07:42:56.420529 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l2kjg" podStartSLOduration=1.732913044 podStartE2EDuration="4.420513978s" podCreationTimestamp="2025-12-12 07:42:52 +0000 UTC" firstStartedPulling="2025-12-12 07:42:53.361751267 +0000 UTC m=+3260.933132536" lastFinishedPulling="2025-12-12 07:42:56.049352201 +0000 UTC m=+3263.620733470" observedRunningTime="2025-12-12 07:42:56.419004101 +0000 UTC m=+3263.990385370" watchObservedRunningTime="2025-12-12 07:42:56.420513978 +0000 UTC m=+3263.991895247" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.373294 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.373729 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.413767 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.459088 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.459151 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.486916 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:00 crc kubenswrapper[4867]: I1212 07:43:00.498145 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:01 crc kubenswrapper[4867]: I1212 07:43:01.505518 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.221435 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.459651 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m5pv5" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="registry-server" containerID="cri-o://5f16d02d6b2458a4101c7fe14b4488f271370c717e7fdb097938cad6d6e907d4" gracePeriod=2 Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.610373 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.610545 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.656084 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:02 crc kubenswrapper[4867]: I1212 07:43:02.823159 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:43:03 crc kubenswrapper[4867]: I1212 07:43:03.467994 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2f9h2" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="registry-server" containerID="cri-o://c88a697cad0acfe56b8081e914ee12740f007aabe67852d17711460ca4c28a75" gracePeriod=2 Dec 12 07:43:03 crc kubenswrapper[4867]: I1212 07:43:03.514851 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.479443 4867 generic.go:334] "Generic (PLEG): container finished" podID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerID="c88a697cad0acfe56b8081e914ee12740f007aabe67852d17711460ca4c28a75" exitCode=0 Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.479643 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerDied","Data":"c88a697cad0acfe56b8081e914ee12740f007aabe67852d17711460ca4c28a75"} Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.481323 4867 generic.go:334] "Generic (PLEG): container finished" podID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerID="5f16d02d6b2458a4101c7fe14b4488f271370c717e7fdb097938cad6d6e907d4" exitCode=0 Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.482062 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerDied","Data":"5f16d02d6b2458a4101c7fe14b4488f271370c717e7fdb097938cad6d6e907d4"} Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.564894 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.669262 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content\") pod \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.669338 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26rz9\" (UniqueName: \"kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9\") pod \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.669382 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities\") pod \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\" (UID: \"ee29e2cb-1234-4ed6-9dc1-790d7fce154e\") " Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.671029 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities" (OuterVolumeSpecName: "utilities") pod "ee29e2cb-1234-4ed6-9dc1-790d7fce154e" (UID: "ee29e2cb-1234-4ed6-9dc1-790d7fce154e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.676243 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9" (OuterVolumeSpecName: "kube-api-access-26rz9") pod "ee29e2cb-1234-4ed6-9dc1-790d7fce154e" (UID: "ee29e2cb-1234-4ed6-9dc1-790d7fce154e"). InnerVolumeSpecName "kube-api-access-26rz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.734840 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee29e2cb-1234-4ed6-9dc1-790d7fce154e" (UID: "ee29e2cb-1234-4ed6-9dc1-790d7fce154e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.771324 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.771360 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.771376 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26rz9\" (UniqueName: \"kubernetes.io/projected/ee29e2cb-1234-4ed6-9dc1-790d7fce154e-kube-api-access-26rz9\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.838044 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:43:04 crc kubenswrapper[4867]: E1212 07:43:04.838589 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:43:04 crc kubenswrapper[4867]: I1212 07:43:04.896821 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.074353 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content\") pod \"496cb354-7260-4c20-8e90-6a0ff6ba944c\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.074429 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities\") pod \"496cb354-7260-4c20-8e90-6a0ff6ba944c\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.074464 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdvt2\" (UniqueName: \"kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2\") pod \"496cb354-7260-4c20-8e90-6a0ff6ba944c\" (UID: \"496cb354-7260-4c20-8e90-6a0ff6ba944c\") " Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.075533 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities" (OuterVolumeSpecName: "utilities") pod "496cb354-7260-4c20-8e90-6a0ff6ba944c" (UID: "496cb354-7260-4c20-8e90-6a0ff6ba944c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.078010 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2" (OuterVolumeSpecName: "kube-api-access-wdvt2") pod "496cb354-7260-4c20-8e90-6a0ff6ba944c" (UID: "496cb354-7260-4c20-8e90-6a0ff6ba944c"). InnerVolumeSpecName "kube-api-access-wdvt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.125107 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "496cb354-7260-4c20-8e90-6a0ff6ba944c" (UID: "496cb354-7260-4c20-8e90-6a0ff6ba944c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.175923 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.175970 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/496cb354-7260-4c20-8e90-6a0ff6ba944c-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.175984 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdvt2\" (UniqueName: \"kubernetes.io/projected/496cb354-7260-4c20-8e90-6a0ff6ba944c-kube-api-access-wdvt2\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.219045 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.489805 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2f9h2" event={"ID":"ee29e2cb-1234-4ed6-9dc1-790d7fce154e","Type":"ContainerDied","Data":"5859e142b55ee43c5a44a7d4b87012c75572658d91d156beebb8d70f07342579"} Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.489849 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2f9h2" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.489868 4867 scope.go:117] "RemoveContainer" containerID="c88a697cad0acfe56b8081e914ee12740f007aabe67852d17711460ca4c28a75" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.492328 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5pv5" event={"ID":"496cb354-7260-4c20-8e90-6a0ff6ba944c","Type":"ContainerDied","Data":"40176b5af1795b6cb09d3fd5f584187027b90d3359dcbc0ab8141d1123c1eccd"} Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.492362 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5pv5" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.519932 4867 scope.go:117] "RemoveContainer" containerID="c0759090b69cee5eb37c90bcc1c2ca426126efbbc68c5c8157716687b86685ea" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.519935 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.526109 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2f9h2"] Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.544646 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.544699 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m5pv5"] Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.555951 4867 scope.go:117] "RemoveContainer" containerID="2ecb2ac2b0aa08f42f93e719213b99365a0ae6445c9fe67fdc08e3df9796ce1d" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.577507 4867 scope.go:117] "RemoveContainer" containerID="5f16d02d6b2458a4101c7fe14b4488f271370c717e7fdb097938cad6d6e907d4" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.614288 4867 scope.go:117] "RemoveContainer" containerID="77c07622cb43f89f8b3ba9ec5997f9fd2b8eab6d645641a570e01bb02f3b9146" Dec 12 07:43:05 crc kubenswrapper[4867]: I1212 07:43:05.630392 4867 scope.go:117] "RemoveContainer" containerID="4f279079f7be1f15dc974fb0e77e5a39de6228a97757b9f32844d543887955c4" Dec 12 07:43:06 crc kubenswrapper[4867]: I1212 07:43:06.508258 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l2kjg" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="registry-server" containerID="cri-o://7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2" gracePeriod=2 Dec 12 07:43:06 crc kubenswrapper[4867]: I1212 07:43:06.848820 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" path="/var/lib/kubelet/pods/496cb354-7260-4c20-8e90-6a0ff6ba944c/volumes" Dec 12 07:43:06 crc kubenswrapper[4867]: I1212 07:43:06.850100 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" path="/var/lib/kubelet/pods/ee29e2cb-1234-4ed6-9dc1-790d7fce154e/volumes" Dec 12 07:43:06 crc kubenswrapper[4867]: I1212 07:43:06.855849 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.002356 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq65c\" (UniqueName: \"kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c\") pod \"66f01148-89f0-4473-96e4-ca3050892aab\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.002409 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content\") pod \"66f01148-89f0-4473-96e4-ca3050892aab\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.002429 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities\") pod \"66f01148-89f0-4473-96e4-ca3050892aab\" (UID: \"66f01148-89f0-4473-96e4-ca3050892aab\") " Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.003506 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities" (OuterVolumeSpecName: "utilities") pod "66f01148-89f0-4473-96e4-ca3050892aab" (UID: "66f01148-89f0-4473-96e4-ca3050892aab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.003816 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.008186 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c" (OuterVolumeSpecName: "kube-api-access-lq65c") pod "66f01148-89f0-4473-96e4-ca3050892aab" (UID: "66f01148-89f0-4473-96e4-ca3050892aab"). InnerVolumeSpecName "kube-api-access-lq65c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.023623 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66f01148-89f0-4473-96e4-ca3050892aab" (UID: "66f01148-89f0-4473-96e4-ca3050892aab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.105466 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq65c\" (UniqueName: \"kubernetes.io/projected/66f01148-89f0-4473-96e4-ca3050892aab-kube-api-access-lq65c\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.105520 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66f01148-89f0-4473-96e4-ca3050892aab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.522522 4867 generic.go:334] "Generic (PLEG): container finished" podID="66f01148-89f0-4473-96e4-ca3050892aab" containerID="7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2" exitCode=0 Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.522584 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerDied","Data":"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2"} Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.522633 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2kjg" event={"ID":"66f01148-89f0-4473-96e4-ca3050892aab","Type":"ContainerDied","Data":"5a1381637c1030ba53e5ee68d19d650f8a236d5e38ef89cfac1b4d9e146042b6"} Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.522643 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2kjg" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.522656 4867 scope.go:117] "RemoveContainer" containerID="7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.549409 4867 scope.go:117] "RemoveContainer" containerID="348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.561263 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.571256 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2kjg"] Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.578463 4867 scope.go:117] "RemoveContainer" containerID="0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.598867 4867 scope.go:117] "RemoveContainer" containerID="7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2" Dec 12 07:43:07 crc kubenswrapper[4867]: E1212 07:43:07.599703 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2\": container with ID starting with 7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2 not found: ID does not exist" containerID="7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.599897 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2"} err="failed to get container status \"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2\": rpc error: code = NotFound desc = could not find container \"7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2\": container with ID starting with 7b7ce2f21f1f04c52fdf1c27a31d652d347a4ccc8b928d5356bbbe4ff158d7e2 not found: ID does not exist" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.600021 4867 scope.go:117] "RemoveContainer" containerID="348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57" Dec 12 07:43:07 crc kubenswrapper[4867]: E1212 07:43:07.600788 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57\": container with ID starting with 348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57 not found: ID does not exist" containerID="348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.600834 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57"} err="failed to get container status \"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57\": rpc error: code = NotFound desc = could not find container \"348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57\": container with ID starting with 348de2d5b6d5ec9b1100c642cc0a0f710ddd1d12f2a39179de35193fe0754e57 not found: ID does not exist" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.600862 4867 scope.go:117] "RemoveContainer" containerID="0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf" Dec 12 07:43:07 crc kubenswrapper[4867]: E1212 07:43:07.601253 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf\": container with ID starting with 0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf not found: ID does not exist" containerID="0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf" Dec 12 07:43:07 crc kubenswrapper[4867]: I1212 07:43:07.601303 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf"} err="failed to get container status \"0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf\": rpc error: code = NotFound desc = could not find container \"0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf\": container with ID starting with 0b653f2bb8ed8746109e9024c86f4e8aad2eceb99c58811070741c8cb25ab4cf not found: ID does not exist" Dec 12 07:43:08 crc kubenswrapper[4867]: I1212 07:43:08.848500 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66f01148-89f0-4473-96e4-ca3050892aab" path="/var/lib/kubelet/pods/66f01148-89f0-4473-96e4-ca3050892aab/volumes" Dec 12 07:43:19 crc kubenswrapper[4867]: I1212 07:43:19.838807 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:43:19 crc kubenswrapper[4867]: E1212 07:43:19.839742 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:43:30 crc kubenswrapper[4867]: I1212 07:43:30.837896 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:43:30 crc kubenswrapper[4867]: E1212 07:43:30.838657 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:43:44 crc kubenswrapper[4867]: I1212 07:43:44.838597 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:43:44 crc kubenswrapper[4867]: E1212 07:43:44.839403 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:43:57 crc kubenswrapper[4867]: I1212 07:43:57.838339 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:43:57 crc kubenswrapper[4867]: E1212 07:43:57.839215 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:44:11 crc kubenswrapper[4867]: I1212 07:44:11.838842 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:44:11 crc kubenswrapper[4867]: E1212 07:44:11.839960 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:44:23 crc kubenswrapper[4867]: I1212 07:44:23.837860 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:44:23 crc kubenswrapper[4867]: E1212 07:44:23.838585 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:44:34 crc kubenswrapper[4867]: I1212 07:44:34.837778 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:44:34 crc kubenswrapper[4867]: E1212 07:44:34.838383 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.368390 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.371337 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.371436 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.371512 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.371579 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.371661 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.371726 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.371808 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.371875 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.371953 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372026 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.372105 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372179 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.372282 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372362 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.372460 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372530 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="extract-content" Dec 12 07:44:44 crc kubenswrapper[4867]: E1212 07:44:44.372608 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372685 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="extract-utilities" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.372955 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee29e2cb-1234-4ed6-9dc1-790d7fce154e" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.373055 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f01148-89f0-4473-96e4-ca3050892aab" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.373133 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="496cb354-7260-4c20-8e90-6a0ff6ba944c" containerName="registry-server" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.374497 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.374755 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.376655 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lqrv\" (UniqueName: \"kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.376817 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.376876 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.479662 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.479733 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.479784 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lqrv\" (UniqueName: \"kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.480454 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.480602 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.509588 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lqrv\" (UniqueName: \"kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv\") pod \"redhat-operators-pn2lh\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:44 crc kubenswrapper[4867]: I1212 07:44:44.704887 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:45 crc kubenswrapper[4867]: I1212 07:44:45.177797 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:44:45 crc kubenswrapper[4867]: I1212 07:44:45.187852 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerStarted","Data":"7ef23a0573dfd5c59dab50d1ee38634c7c668e1b1adac008a91b4b4a457af56b"} Dec 12 07:44:46 crc kubenswrapper[4867]: I1212 07:44:46.196546 4867 generic.go:334] "Generic (PLEG): container finished" podID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerID="a63840376ef1f4630085e9521f81ec812cda72c85ee4a6e349d6d89f4ed863cf" exitCode=0 Dec 12 07:44:46 crc kubenswrapper[4867]: I1212 07:44:46.196657 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerDied","Data":"a63840376ef1f4630085e9521f81ec812cda72c85ee4a6e349d6d89f4ed863cf"} Dec 12 07:44:47 crc kubenswrapper[4867]: I1212 07:44:47.208322 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerStarted","Data":"d162fe67f3e27e2f78f531f494b0ec0fe32be20c65cec2d0f9d9877c00e7c27b"} Dec 12 07:44:48 crc kubenswrapper[4867]: I1212 07:44:48.216925 4867 generic.go:334] "Generic (PLEG): container finished" podID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerID="d162fe67f3e27e2f78f531f494b0ec0fe32be20c65cec2d0f9d9877c00e7c27b" exitCode=0 Dec 12 07:44:48 crc kubenswrapper[4867]: I1212 07:44:48.216979 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerDied","Data":"d162fe67f3e27e2f78f531f494b0ec0fe32be20c65cec2d0f9d9877c00e7c27b"} Dec 12 07:44:49 crc kubenswrapper[4867]: I1212 07:44:49.228058 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerStarted","Data":"e790225238697b521175d508ffce6f2614db02ef8cc1a4fc49d5b6f2d9cf8f93"} Dec 12 07:44:49 crc kubenswrapper[4867]: I1212 07:44:49.838007 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:44:49 crc kubenswrapper[4867]: E1212 07:44:49.838238 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:44:54 crc kubenswrapper[4867]: I1212 07:44:54.705906 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:54 crc kubenswrapper[4867]: I1212 07:44:54.706529 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:54 crc kubenswrapper[4867]: I1212 07:44:54.758590 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:54 crc kubenswrapper[4867]: I1212 07:44:54.783337 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pn2lh" podStartSLOduration=8.227679371 podStartE2EDuration="10.783303758s" podCreationTimestamp="2025-12-12 07:44:44 +0000 UTC" firstStartedPulling="2025-12-12 07:44:46.197964884 +0000 UTC m=+3373.769346153" lastFinishedPulling="2025-12-12 07:44:48.753589281 +0000 UTC m=+3376.324970540" observedRunningTime="2025-12-12 07:44:49.249909838 +0000 UTC m=+3376.821291107" watchObservedRunningTime="2025-12-12 07:44:54.783303758 +0000 UTC m=+3382.354685027" Dec 12 07:44:55 crc kubenswrapper[4867]: I1212 07:44:55.308354 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:44:55 crc kubenswrapper[4867]: I1212 07:44:55.360852 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:44:57 crc kubenswrapper[4867]: I1212 07:44:57.280125 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pn2lh" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="registry-server" containerID="cri-o://e790225238697b521175d508ffce6f2614db02ef8cc1a4fc49d5b6f2d9cf8f93" gracePeriod=2 Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.142656 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg"] Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.144707 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.147680 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.155329 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.156879 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg"] Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.201773 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjqf\" (UniqueName: \"kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.201900 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.202110 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.299172 4867 generic.go:334] "Generic (PLEG): container finished" podID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerID="e790225238697b521175d508ffce6f2614db02ef8cc1a4fc49d5b6f2d9cf8f93" exitCode=0 Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.299238 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerDied","Data":"e790225238697b521175d508ffce6f2614db02ef8cc1a4fc49d5b6f2d9cf8f93"} Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.303016 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjqf\" (UniqueName: \"kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.303079 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.303301 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.305046 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.309851 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.320916 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjqf\" (UniqueName: \"kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf\") pod \"collect-profiles-29425425-rpdcg\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.532184 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:00 crc kubenswrapper[4867]: I1212 07:45:00.957971 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg"] Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.309258 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" event={"ID":"8d1526b2-f850-4458-b1fa-b6f094eb2c26","Type":"ContainerStarted","Data":"5aaa22c4ccd909c7bbaf458dcb91e8c8a8c6d6955ffe7cfc8b10805fdc09b52c"} Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.800714 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.827101 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities\") pod \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.827277 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content\") pod \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.827454 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lqrv\" (UniqueName: \"kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv\") pod \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\" (UID: \"b659aaa5-89cd-4721-b005-f4e61eb3af9a\") " Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.827990 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities" (OuterVolumeSpecName: "utilities") pod "b659aaa5-89cd-4721-b005-f4e61eb3af9a" (UID: "b659aaa5-89cd-4721-b005-f4e61eb3af9a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.833892 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv" (OuterVolumeSpecName: "kube-api-access-5lqrv") pod "b659aaa5-89cd-4721-b005-f4e61eb3af9a" (UID: "b659aaa5-89cd-4721-b005-f4e61eb3af9a"). InnerVolumeSpecName "kube-api-access-5lqrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.838858 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:45:01 crc kubenswrapper[4867]: E1212 07:45:01.839214 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.929945 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lqrv\" (UniqueName: \"kubernetes.io/projected/b659aaa5-89cd-4721-b005-f4e61eb3af9a-kube-api-access-5lqrv\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.929991 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:01 crc kubenswrapper[4867]: I1212 07:45:01.955990 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b659aaa5-89cd-4721-b005-f4e61eb3af9a" (UID: "b659aaa5-89cd-4721-b005-f4e61eb3af9a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.031767 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b659aaa5-89cd-4721-b005-f4e61eb3af9a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.315695 4867 generic.go:334] "Generic (PLEG): container finished" podID="8d1526b2-f850-4458-b1fa-b6f094eb2c26" containerID="ab8dd21f54dce96713b525c86ade91bc53a534754bdb926c40507f9a80a0f5d5" exitCode=0 Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.315767 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" event={"ID":"8d1526b2-f850-4458-b1fa-b6f094eb2c26","Type":"ContainerDied","Data":"ab8dd21f54dce96713b525c86ade91bc53a534754bdb926c40507f9a80a0f5d5"} Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.317875 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pn2lh" event={"ID":"b659aaa5-89cd-4721-b005-f4e61eb3af9a","Type":"ContainerDied","Data":"7ef23a0573dfd5c59dab50d1ee38634c7c668e1b1adac008a91b4b4a457af56b"} Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.317928 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pn2lh" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.317930 4867 scope.go:117] "RemoveContainer" containerID="e790225238697b521175d508ffce6f2614db02ef8cc1a4fc49d5b6f2d9cf8f93" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.342426 4867 scope.go:117] "RemoveContainer" containerID="d162fe67f3e27e2f78f531f494b0ec0fe32be20c65cec2d0f9d9877c00e7c27b" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.361040 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.366692 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pn2lh"] Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.389717 4867 scope.go:117] "RemoveContainer" containerID="a63840376ef1f4630085e9521f81ec812cda72c85ee4a6e349d6d89f4ed863cf" Dec 12 07:45:02 crc kubenswrapper[4867]: I1212 07:45:02.849792 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" path="/var/lib/kubelet/pods/b659aaa5-89cd-4721-b005-f4e61eb3af9a/volumes" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.569955 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.655670 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume\") pod \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.655747 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume\") pod \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.655886 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsjqf\" (UniqueName: \"kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf\") pod \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\" (UID: \"8d1526b2-f850-4458-b1fa-b6f094eb2c26\") " Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.656437 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d1526b2-f850-4458-b1fa-b6f094eb2c26" (UID: "8d1526b2-f850-4458-b1fa-b6f094eb2c26"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.661118 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d1526b2-f850-4458-b1fa-b6f094eb2c26" (UID: "8d1526b2-f850-4458-b1fa-b6f094eb2c26"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.661140 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf" (OuterVolumeSpecName: "kube-api-access-dsjqf") pod "8d1526b2-f850-4458-b1fa-b6f094eb2c26" (UID: "8d1526b2-f850-4458-b1fa-b6f094eb2c26"). InnerVolumeSpecName "kube-api-access-dsjqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.757701 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d1526b2-f850-4458-b1fa-b6f094eb2c26-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.757744 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d1526b2-f850-4458-b1fa-b6f094eb2c26-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:03 crc kubenswrapper[4867]: I1212 07:45:03.757758 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsjqf\" (UniqueName: \"kubernetes.io/projected/8d1526b2-f850-4458-b1fa-b6f094eb2c26-kube-api-access-dsjqf\") on node \"crc\" DevicePath \"\"" Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.364831 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" event={"ID":"8d1526b2-f850-4458-b1fa-b6f094eb2c26","Type":"ContainerDied","Data":"5aaa22c4ccd909c7bbaf458dcb91e8c8a8c6d6955ffe7cfc8b10805fdc09b52c"} Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.364882 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aaa22c4ccd909c7bbaf458dcb91e8c8a8c6d6955ffe7cfc8b10805fdc09b52c" Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.364893 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg" Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.634619 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk"] Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.641815 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425380-d6wpk"] Dec 12 07:45:04 crc kubenswrapper[4867]: I1212 07:45:04.880886 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c40f17a4-60eb-4246-aa3f-aa2cf573542d" path="/var/lib/kubelet/pods/c40f17a4-60eb-4246-aa3f-aa2cf573542d/volumes" Dec 12 07:45:15 crc kubenswrapper[4867]: I1212 07:45:15.838109 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:45:15 crc kubenswrapper[4867]: E1212 07:45:15.839032 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:45:29 crc kubenswrapper[4867]: I1212 07:45:29.838393 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:45:30 crc kubenswrapper[4867]: I1212 07:45:30.558369 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25"} Dec 12 07:45:43 crc kubenswrapper[4867]: I1212 07:45:43.112068 4867 scope.go:117] "RemoveContainer" containerID="86e82217785c1207c428940654715be7d9cee5dd0cfca8e19105f8e3ebbcc836" Dec 12 07:47:58 crc kubenswrapper[4867]: I1212 07:47:58.989169 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:47:58 crc kubenswrapper[4867]: I1212 07:47:58.989832 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:48:28 crc kubenswrapper[4867]: I1212 07:48:28.989127 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:48:28 crc kubenswrapper[4867]: I1212 07:48:28.989808 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:48:58 crc kubenswrapper[4867]: I1212 07:48:58.989033 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:48:58 crc kubenswrapper[4867]: I1212 07:48:58.989600 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:48:58 crc kubenswrapper[4867]: I1212 07:48:58.989639 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:48:58 crc kubenswrapper[4867]: I1212 07:48:58.990291 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:48:58 crc kubenswrapper[4867]: I1212 07:48:58.990346 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25" gracePeriod=600 Dec 12 07:49:00 crc kubenswrapper[4867]: I1212 07:49:00.082423 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25" exitCode=0 Dec 12 07:49:00 crc kubenswrapper[4867]: I1212 07:49:00.082498 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25"} Dec 12 07:49:00 crc kubenswrapper[4867]: I1212 07:49:00.082997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb"} Dec 12 07:49:00 crc kubenswrapper[4867]: I1212 07:49:00.083022 4867 scope.go:117] "RemoveContainer" containerID="beddaa203bb0d09901f29abfa1b3e6e3494c0b140d3075a1cd895c677354145b" Dec 12 07:51:28 crc kubenswrapper[4867]: I1212 07:51:28.988498 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:51:28 crc kubenswrapper[4867]: I1212 07:51:28.989042 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:51:58 crc kubenswrapper[4867]: I1212 07:51:58.988877 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:51:58 crc kubenswrapper[4867]: I1212 07:51:58.989452 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:52:28 crc kubenswrapper[4867]: I1212 07:52:28.989065 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:52:28 crc kubenswrapper[4867]: I1212 07:52:28.990069 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 07:52:28 crc kubenswrapper[4867]: I1212 07:52:28.990127 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 07:52:28 crc kubenswrapper[4867]: I1212 07:52:28.991002 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 07:52:28 crc kubenswrapper[4867]: I1212 07:52:28.991065 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" gracePeriod=600 Dec 12 07:52:29 crc kubenswrapper[4867]: E1212 07:52:29.265686 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:52:29 crc kubenswrapper[4867]: I1212 07:52:29.663262 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" exitCode=0 Dec 12 07:52:29 crc kubenswrapper[4867]: I1212 07:52:29.663355 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb"} Dec 12 07:52:29 crc kubenswrapper[4867]: I1212 07:52:29.663706 4867 scope.go:117] "RemoveContainer" containerID="82194d86ceb7f37e3df9e0f0cdedf9cdcbb0b087ec6aa28fc424ed4713f99c25" Dec 12 07:52:29 crc kubenswrapper[4867]: I1212 07:52:29.664401 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:52:29 crc kubenswrapper[4867]: E1212 07:52:29.664665 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:52:40 crc kubenswrapper[4867]: I1212 07:52:40.838520 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:52:40 crc kubenswrapper[4867]: E1212 07:52:40.839527 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:52:54 crc kubenswrapper[4867]: I1212 07:52:54.837894 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:52:54 crc kubenswrapper[4867]: E1212 07:52:54.838594 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:05 crc kubenswrapper[4867]: I1212 07:53:05.838114 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:53:05 crc kubenswrapper[4867]: E1212 07:53:05.838920 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:20 crc kubenswrapper[4867]: I1212 07:53:20.838305 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:53:20 crc kubenswrapper[4867]: E1212 07:53:20.839108 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.005539 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:25 crc kubenswrapper[4867]: E1212 07:53:25.006583 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1526b2-f850-4458-b1fa-b6f094eb2c26" containerName="collect-profiles" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006598 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1526b2-f850-4458-b1fa-b6f094eb2c26" containerName="collect-profiles" Dec 12 07:53:25 crc kubenswrapper[4867]: E1212 07:53:25.006613 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="registry-server" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006618 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="registry-server" Dec 12 07:53:25 crc kubenswrapper[4867]: E1212 07:53:25.006643 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="extract-content" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006650 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="extract-content" Dec 12 07:53:25 crc kubenswrapper[4867]: E1212 07:53:25.006658 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="extract-utilities" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006665 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="extract-utilities" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006847 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b659aaa5-89cd-4721-b005-f4e61eb3af9a" containerName="registry-server" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.006871 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1526b2-f850-4458-b1fa-b6f094eb2c26" containerName="collect-profiles" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.009202 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.022090 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.172541 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.172607 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq48g\" (UniqueName: \"kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.172698 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.273558 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.273682 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.273705 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq48g\" (UniqueName: \"kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.274490 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.274510 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.605295 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq48g\" (UniqueName: \"kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g\") pod \"certified-operators-w5lnk\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:25 crc kubenswrapper[4867]: I1212 07:53:25.670522 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.084832 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.794378 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2rmpq"] Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.796382 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.804768 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2rmpq"] Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.895353 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-utilities\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.895694 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwwqg\" (UniqueName: \"kubernetes.io/projected/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-kube-api-access-vwwqg\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.895718 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-catalog-content\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.997290 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-utilities\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.997368 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwwqg\" (UniqueName: \"kubernetes.io/projected/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-kube-api-access-vwwqg\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.997395 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-catalog-content\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.997888 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-utilities\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:26 crc kubenswrapper[4867]: I1212 07:53:26.998815 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-catalog-content\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.058769 4867 generic.go:334] "Generic (PLEG): container finished" podID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerID="22dd7499f94d817e004a4d71d458eaeb779fca1227522bb781b0cd29b76b2b3e" exitCode=0 Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.058842 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerDied","Data":"22dd7499f94d817e004a4d71d458eaeb779fca1227522bb781b0cd29b76b2b3e"} Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.058891 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerStarted","Data":"83f512905f3d73cb2d30c66264c09f7eda34307a36cd5c629d05971797885a53"} Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.061879 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.103512 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwwqg\" (UniqueName: \"kubernetes.io/projected/715ca56a-e74a-4ca4-a30c-7779d73b4fb4-kube-api-access-vwwqg\") pod \"community-operators-2rmpq\" (UID: \"715ca56a-e74a-4ca4-a30c-7779d73b4fb4\") " pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.118850 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:27 crc kubenswrapper[4867]: I1212 07:53:27.443244 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2rmpq"] Dec 12 07:53:28 crc kubenswrapper[4867]: I1212 07:53:28.066206 4867 generic.go:334] "Generic (PLEG): container finished" podID="715ca56a-e74a-4ca4-a30c-7779d73b4fb4" containerID="ba623c47e369c5e9f63f827dec26f31de580e3680ba24a89d4a69dbf1ed8a86b" exitCode=0 Dec 12 07:53:28 crc kubenswrapper[4867]: I1212 07:53:28.066290 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rmpq" event={"ID":"715ca56a-e74a-4ca4-a30c-7779d73b4fb4","Type":"ContainerDied","Data":"ba623c47e369c5e9f63f827dec26f31de580e3680ba24a89d4a69dbf1ed8a86b"} Dec 12 07:53:28 crc kubenswrapper[4867]: I1212 07:53:28.066370 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rmpq" event={"ID":"715ca56a-e74a-4ca4-a30c-7779d73b4fb4","Type":"ContainerStarted","Data":"61083820c5fc8d1ac61b5aa2b5ba517c94a829408a2bc619b3dfc7286df83454"} Dec 12 07:53:28 crc kubenswrapper[4867]: I1212 07:53:28.068575 4867 generic.go:334] "Generic (PLEG): container finished" podID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerID="7c5f2762f50c9155ef38180dbd607dbc73a3c49215d18b881da290923091a0d9" exitCode=0 Dec 12 07:53:28 crc kubenswrapper[4867]: I1212 07:53:28.068599 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerDied","Data":"7c5f2762f50c9155ef38180dbd607dbc73a3c49215d18b881da290923091a0d9"} Dec 12 07:53:29 crc kubenswrapper[4867]: I1212 07:53:29.086498 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerStarted","Data":"d910636b4e616cf80822c65ec72b2ffeb714b615d5dd1d4652a8442b2f78ca36"} Dec 12 07:53:29 crc kubenswrapper[4867]: I1212 07:53:29.111742 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w5lnk" podStartSLOduration=3.6539166229999998 podStartE2EDuration="5.111713654s" podCreationTimestamp="2025-12-12 07:53:24 +0000 UTC" firstStartedPulling="2025-12-12 07:53:27.061586956 +0000 UTC m=+3894.632968225" lastFinishedPulling="2025-12-12 07:53:28.519383987 +0000 UTC m=+3896.090765256" observedRunningTime="2025-12-12 07:53:29.102870926 +0000 UTC m=+3896.674252205" watchObservedRunningTime="2025-12-12 07:53:29.111713654 +0000 UTC m=+3896.683094933" Dec 12 07:53:32 crc kubenswrapper[4867]: I1212 07:53:32.125777 4867 generic.go:334] "Generic (PLEG): container finished" podID="715ca56a-e74a-4ca4-a30c-7779d73b4fb4" containerID="5e7fbfa4224eecc2b675c2f0a05ad822012321e36a8ac9363a88eb8de0ed3346" exitCode=0 Dec 12 07:53:32 crc kubenswrapper[4867]: I1212 07:53:32.127462 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rmpq" event={"ID":"715ca56a-e74a-4ca4-a30c-7779d73b4fb4","Type":"ContainerDied","Data":"5e7fbfa4224eecc2b675c2f0a05ad822012321e36a8ac9363a88eb8de0ed3346"} Dec 12 07:53:33 crc kubenswrapper[4867]: I1212 07:53:33.134746 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rmpq" event={"ID":"715ca56a-e74a-4ca4-a30c-7779d73b4fb4","Type":"ContainerStarted","Data":"b3fc8c54247f89720e0eb11d50bec20f4dd9c5ed771ce6b2272d6a54004006a2"} Dec 12 07:53:33 crc kubenswrapper[4867]: I1212 07:53:33.151544 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2rmpq" podStartSLOduration=2.239299388 podStartE2EDuration="7.151525352s" podCreationTimestamp="2025-12-12 07:53:26 +0000 UTC" firstStartedPulling="2025-12-12 07:53:28.06980894 +0000 UTC m=+3895.641190209" lastFinishedPulling="2025-12-12 07:53:32.982034904 +0000 UTC m=+3900.553416173" observedRunningTime="2025-12-12 07:53:33.149179714 +0000 UTC m=+3900.720560993" watchObservedRunningTime="2025-12-12 07:53:33.151525352 +0000 UTC m=+3900.722906621" Dec 12 07:53:34 crc kubenswrapper[4867]: I1212 07:53:34.837837 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:53:34 crc kubenswrapper[4867]: E1212 07:53:34.838467 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:35 crc kubenswrapper[4867]: I1212 07:53:35.671081 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:35 crc kubenswrapper[4867]: I1212 07:53:35.671336 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:35 crc kubenswrapper[4867]: I1212 07:53:35.742035 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:36 crc kubenswrapper[4867]: I1212 07:53:36.204126 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:36 crc kubenswrapper[4867]: I1212 07:53:36.249740 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:37 crc kubenswrapper[4867]: I1212 07:53:37.119333 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:37 crc kubenswrapper[4867]: I1212 07:53:37.120382 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:37 crc kubenswrapper[4867]: I1212 07:53:37.224403 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:38 crc kubenswrapper[4867]: I1212 07:53:38.170985 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w5lnk" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="registry-server" containerID="cri-o://d910636b4e616cf80822c65ec72b2ffeb714b615d5dd1d4652a8442b2f78ca36" gracePeriod=2 Dec 12 07:53:38 crc kubenswrapper[4867]: I1212 07:53:38.242792 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2rmpq" Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.250251 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2rmpq"] Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.388718 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.389020 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hszdl" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="registry-server" containerID="cri-o://437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63" gracePeriod=2 Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.826451 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hszdl" Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.986316 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xqvp\" (UniqueName: \"kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp\") pod \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.986361 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content\") pod \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.986395 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities\") pod \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\" (UID: \"ac8f3914-dcac-4c7d-8fa5-6c7648197f43\") " Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.987347 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities" (OuterVolumeSpecName: "utilities") pod "ac8f3914-dcac-4c7d-8fa5-6c7648197f43" (UID: "ac8f3914-dcac-4c7d-8fa5-6c7648197f43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:53:39 crc kubenswrapper[4867]: I1212 07:53:39.993073 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp" (OuterVolumeSpecName: "kube-api-access-7xqvp") pod "ac8f3914-dcac-4c7d-8fa5-6c7648197f43" (UID: "ac8f3914-dcac-4c7d-8fa5-6c7648197f43"). InnerVolumeSpecName "kube-api-access-7xqvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.038579 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac8f3914-dcac-4c7d-8fa5-6c7648197f43" (UID: "ac8f3914-dcac-4c7d-8fa5-6c7648197f43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.087858 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xqvp\" (UniqueName: \"kubernetes.io/projected/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-kube-api-access-7xqvp\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.087896 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.087906 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac8f3914-dcac-4c7d-8fa5-6c7648197f43-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.185716 4867 generic.go:334] "Generic (PLEG): container finished" podID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerID="437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63" exitCode=0 Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.185770 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hszdl" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.185839 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerDied","Data":"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63"} Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.185916 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hszdl" event={"ID":"ac8f3914-dcac-4c7d-8fa5-6c7648197f43","Type":"ContainerDied","Data":"1275973da752f01890689a8dfac657ecaf3ca55a10ad60265fc6815a8a8cd651"} Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.185950 4867 scope.go:117] "RemoveContainer" containerID="437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.188481 4867 generic.go:334] "Generic (PLEG): container finished" podID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerID="d910636b4e616cf80822c65ec72b2ffeb714b615d5dd1d4652a8442b2f78ca36" exitCode=0 Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.188545 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerDied","Data":"d910636b4e616cf80822c65ec72b2ffeb714b615d5dd1d4652a8442b2f78ca36"} Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.238551 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.243486 4867 scope.go:117] "RemoveContainer" containerID="135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.245468 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hszdl"] Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.269497 4867 scope.go:117] "RemoveContainer" containerID="4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.295255 4867 scope.go:117] "RemoveContainer" containerID="437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63" Dec 12 07:53:40 crc kubenswrapper[4867]: E1212 07:53:40.296934 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63\": container with ID starting with 437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63 not found: ID does not exist" containerID="437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.296992 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63"} err="failed to get container status \"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63\": rpc error: code = NotFound desc = could not find container \"437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63\": container with ID starting with 437c28d27f467f3b40b652270cbd87890263448988b24669eeec44a9acd4fd63 not found: ID does not exist" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.297024 4867 scope.go:117] "RemoveContainer" containerID="135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125" Dec 12 07:53:40 crc kubenswrapper[4867]: E1212 07:53:40.298938 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125\": container with ID starting with 135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125 not found: ID does not exist" containerID="135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.298985 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125"} err="failed to get container status \"135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125\": rpc error: code = NotFound desc = could not find container \"135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125\": container with ID starting with 135a4b841fcce8b6ff78c8d7cd56b70ea44aa6c9fdf3802f0b34593a24ebc125 not found: ID does not exist" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.299016 4867 scope.go:117] "RemoveContainer" containerID="4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049" Dec 12 07:53:40 crc kubenswrapper[4867]: E1212 07:53:40.299920 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049\": container with ID starting with 4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049 not found: ID does not exist" containerID="4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.299949 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049"} err="failed to get container status \"4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049\": rpc error: code = NotFound desc = could not find container \"4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049\": container with ID starting with 4e7492aaa3811dcac9f9bb8783d286ea985a5a1428ee59ecb10f2f91207b0049 not found: ID does not exist" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.378026 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.492708 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content\") pod \"74391237-7334-4ad3-abc9-8d4c036e06bc\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.492819 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq48g\" (UniqueName: \"kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g\") pod \"74391237-7334-4ad3-abc9-8d4c036e06bc\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.492888 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities\") pod \"74391237-7334-4ad3-abc9-8d4c036e06bc\" (UID: \"74391237-7334-4ad3-abc9-8d4c036e06bc\") " Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.493924 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities" (OuterVolumeSpecName: "utilities") pod "74391237-7334-4ad3-abc9-8d4c036e06bc" (UID: "74391237-7334-4ad3-abc9-8d4c036e06bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.498644 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g" (OuterVolumeSpecName: "kube-api-access-hq48g") pod "74391237-7334-4ad3-abc9-8d4c036e06bc" (UID: "74391237-7334-4ad3-abc9-8d4c036e06bc"). InnerVolumeSpecName "kube-api-access-hq48g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.549837 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74391237-7334-4ad3-abc9-8d4c036e06bc" (UID: "74391237-7334-4ad3-abc9-8d4c036e06bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.595049 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.595088 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq48g\" (UniqueName: \"kubernetes.io/projected/74391237-7334-4ad3-abc9-8d4c036e06bc-kube-api-access-hq48g\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.595111 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74391237-7334-4ad3-abc9-8d4c036e06bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:53:40 crc kubenswrapper[4867]: I1212 07:53:40.847787 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" path="/var/lib/kubelet/pods/ac8f3914-dcac-4c7d-8fa5-6c7648197f43/volumes" Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.197698 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w5lnk" event={"ID":"74391237-7334-4ad3-abc9-8d4c036e06bc","Type":"ContainerDied","Data":"83f512905f3d73cb2d30c66264c09f7eda34307a36cd5c629d05971797885a53"} Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.197769 4867 scope.go:117] "RemoveContainer" containerID="d910636b4e616cf80822c65ec72b2ffeb714b615d5dd1d4652a8442b2f78ca36" Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.197708 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w5lnk" Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.220667 4867 scope.go:117] "RemoveContainer" containerID="7c5f2762f50c9155ef38180dbd607dbc73a3c49215d18b881da290923091a0d9" Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.224112 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.231842 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w5lnk"] Dec 12 07:53:41 crc kubenswrapper[4867]: I1212 07:53:41.236284 4867 scope.go:117] "RemoveContainer" containerID="22dd7499f94d817e004a4d71d458eaeb779fca1227522bb781b0cd29b76b2b3e" Dec 12 07:53:42 crc kubenswrapper[4867]: I1212 07:53:42.850321 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" path="/var/lib/kubelet/pods/74391237-7334-4ad3-abc9-8d4c036e06bc/volumes" Dec 12 07:53:46 crc kubenswrapper[4867]: I1212 07:53:46.837660 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:53:46 crc kubenswrapper[4867]: E1212 07:53:46.838164 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595102 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595533 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595550 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595562 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="extract-content" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595570 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="extract-content" Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595581 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="extract-utilities" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595590 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="extract-utilities" Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595616 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="extract-content" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595623 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="extract-content" Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595633 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="extract-utilities" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595640 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="extract-utilities" Dec 12 07:53:48 crc kubenswrapper[4867]: E1212 07:53:48.595656 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595663 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595858 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="74391237-7334-4ad3-abc9-8d4c036e06bc" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.595876 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8f3914-dcac-4c7d-8fa5-6c7648197f43" containerName="registry-server" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.597067 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.616695 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.701560 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.701659 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.701745 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxhp6\" (UniqueName: \"kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.803134 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxhp6\" (UniqueName: \"kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.803199 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.803279 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.803712 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.804218 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.823190 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxhp6\" (UniqueName: \"kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6\") pod \"redhat-marketplace-6tfkj\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:48 crc kubenswrapper[4867]: I1212 07:53:48.917434 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:49 crc kubenswrapper[4867]: I1212 07:53:49.378137 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:53:49 crc kubenswrapper[4867]: W1212 07:53:49.381820 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod001da497_b1b9_4a18_a03a_bd08ba3e6499.slice/crio-ad335df952f15b2d2ff8d8b783464bda1b7e1a22bc49a0023f802ef98fec0a77 WatchSource:0}: Error finding container ad335df952f15b2d2ff8d8b783464bda1b7e1a22bc49a0023f802ef98fec0a77: Status 404 returned error can't find the container with id ad335df952f15b2d2ff8d8b783464bda1b7e1a22bc49a0023f802ef98fec0a77 Dec 12 07:53:50 crc kubenswrapper[4867]: I1212 07:53:50.267167 4867 generic.go:334] "Generic (PLEG): container finished" podID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerID="e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5" exitCode=0 Dec 12 07:53:50 crc kubenswrapper[4867]: I1212 07:53:50.267263 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerDied","Data":"e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5"} Dec 12 07:53:50 crc kubenswrapper[4867]: I1212 07:53:50.267716 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerStarted","Data":"ad335df952f15b2d2ff8d8b783464bda1b7e1a22bc49a0023f802ef98fec0a77"} Dec 12 07:53:51 crc kubenswrapper[4867]: I1212 07:53:51.276458 4867 generic.go:334] "Generic (PLEG): container finished" podID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerID="d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2" exitCode=0 Dec 12 07:53:51 crc kubenswrapper[4867]: I1212 07:53:51.276735 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerDied","Data":"d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2"} Dec 12 07:53:53 crc kubenswrapper[4867]: I1212 07:53:53.292301 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerStarted","Data":"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f"} Dec 12 07:53:53 crc kubenswrapper[4867]: I1212 07:53:53.315186 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6tfkj" podStartSLOduration=3.396823027 podStartE2EDuration="5.315166023s" podCreationTimestamp="2025-12-12 07:53:48 +0000 UTC" firstStartedPulling="2025-12-12 07:53:50.270151679 +0000 UTC m=+3917.841532948" lastFinishedPulling="2025-12-12 07:53:52.188494685 +0000 UTC m=+3919.759875944" observedRunningTime="2025-12-12 07:53:53.309076854 +0000 UTC m=+3920.880458143" watchObservedRunningTime="2025-12-12 07:53:53.315166023 +0000 UTC m=+3920.886547292" Dec 12 07:53:58 crc kubenswrapper[4867]: I1212 07:53:58.838293 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:53:58 crc kubenswrapper[4867]: E1212 07:53:58.839134 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:53:58 crc kubenswrapper[4867]: I1212 07:53:58.918086 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:58 crc kubenswrapper[4867]: I1212 07:53:58.918165 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:58 crc kubenswrapper[4867]: I1212 07:53:58.960098 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:59 crc kubenswrapper[4867]: I1212 07:53:59.371408 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:53:59 crc kubenswrapper[4867]: I1212 07:53:59.412644 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:54:01 crc kubenswrapper[4867]: I1212 07:54:01.347075 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6tfkj" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="registry-server" containerID="cri-o://c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f" gracePeriod=2 Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.280077 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.355009 4867 generic.go:334] "Generic (PLEG): container finished" podID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerID="c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f" exitCode=0 Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.355053 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerDied","Data":"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f"} Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.355073 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tfkj" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.355085 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tfkj" event={"ID":"001da497-b1b9-4a18-a03a-bd08ba3e6499","Type":"ContainerDied","Data":"ad335df952f15b2d2ff8d8b783464bda1b7e1a22bc49a0023f802ef98fec0a77"} Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.355104 4867 scope.go:117] "RemoveContainer" containerID="c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.369766 4867 scope.go:117] "RemoveContainer" containerID="d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.383180 4867 scope.go:117] "RemoveContainer" containerID="e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.395970 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content\") pod \"001da497-b1b9-4a18-a03a-bd08ba3e6499\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.396069 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxhp6\" (UniqueName: \"kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6\") pod \"001da497-b1b9-4a18-a03a-bd08ba3e6499\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.396331 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities\") pod \"001da497-b1b9-4a18-a03a-bd08ba3e6499\" (UID: \"001da497-b1b9-4a18-a03a-bd08ba3e6499\") " Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.397429 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities" (OuterVolumeSpecName: "utilities") pod "001da497-b1b9-4a18-a03a-bd08ba3e6499" (UID: "001da497-b1b9-4a18-a03a-bd08ba3e6499"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.403023 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6" (OuterVolumeSpecName: "kube-api-access-wxhp6") pod "001da497-b1b9-4a18-a03a-bd08ba3e6499" (UID: "001da497-b1b9-4a18-a03a-bd08ba3e6499"). InnerVolumeSpecName "kube-api-access-wxhp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.414651 4867 scope.go:117] "RemoveContainer" containerID="c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f" Dec 12 07:54:02 crc kubenswrapper[4867]: E1212 07:54:02.415028 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f\": container with ID starting with c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f not found: ID does not exist" containerID="c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.415067 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f"} err="failed to get container status \"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f\": rpc error: code = NotFound desc = could not find container \"c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f\": container with ID starting with c3bde69e9921ca2960ca496a6f42630b0d47c63e28a8e6b7d1124b2cd6d2d49f not found: ID does not exist" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.415090 4867 scope.go:117] "RemoveContainer" containerID="d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2" Dec 12 07:54:02 crc kubenswrapper[4867]: E1212 07:54:02.415328 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2\": container with ID starting with d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2 not found: ID does not exist" containerID="d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.415350 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2"} err="failed to get container status \"d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2\": rpc error: code = NotFound desc = could not find container \"d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2\": container with ID starting with d8519825e2adff7a266b78850c4e0ea030694320c5e9949a2d01e541273e37f2 not found: ID does not exist" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.415363 4867 scope.go:117] "RemoveContainer" containerID="e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5" Dec 12 07:54:02 crc kubenswrapper[4867]: E1212 07:54:02.415811 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5\": container with ID starting with e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5 not found: ID does not exist" containerID="e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.415933 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5"} err="failed to get container status \"e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5\": rpc error: code = NotFound desc = could not find container \"e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5\": container with ID starting with e925e460bdbbd673f43173b42dc8874da3ec3d99755a44fd2e58e88d211635b5 not found: ID does not exist" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.418784 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "001da497-b1b9-4a18-a03a-bd08ba3e6499" (UID: "001da497-b1b9-4a18-a03a-bd08ba3e6499"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.497823 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.497873 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxhp6\" (UniqueName: \"kubernetes.io/projected/001da497-b1b9-4a18-a03a-bd08ba3e6499-kube-api-access-wxhp6\") on node \"crc\" DevicePath \"\"" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.497890 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/001da497-b1b9-4a18-a03a-bd08ba3e6499-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.686288 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.694827 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tfkj"] Dec 12 07:54:02 crc kubenswrapper[4867]: I1212 07:54:02.847903 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" path="/var/lib/kubelet/pods/001da497-b1b9-4a18-a03a-bd08ba3e6499/volumes" Dec 12 07:54:11 crc kubenswrapper[4867]: I1212 07:54:11.837949 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:54:11 crc kubenswrapper[4867]: E1212 07:54:11.838525 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:54:26 crc kubenswrapper[4867]: I1212 07:54:26.838745 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:54:26 crc kubenswrapper[4867]: E1212 07:54:26.839537 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:54:41 crc kubenswrapper[4867]: I1212 07:54:41.838040 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:54:41 crc kubenswrapper[4867]: E1212 07:54:41.838854 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:54:56 crc kubenswrapper[4867]: I1212 07:54:56.841132 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:54:56 crc kubenswrapper[4867]: E1212 07:54:56.841864 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.816532 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:06 crc kubenswrapper[4867]: E1212 07:55:06.817298 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="extract-content" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.817314 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="extract-content" Dec 12 07:55:06 crc kubenswrapper[4867]: E1212 07:55:06.817329 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="registry-server" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.817336 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="registry-server" Dec 12 07:55:06 crc kubenswrapper[4867]: E1212 07:55:06.817353 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="extract-utilities" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.817361 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="extract-utilities" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.817502 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="001da497-b1b9-4a18-a03a-bd08ba3e6499" containerName="registry-server" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.818599 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.828467 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.923690 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.924035 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl4c4\" (UniqueName: \"kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:06 crc kubenswrapper[4867]: I1212 07:55:06.924328 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.025682 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl4c4\" (UniqueName: \"kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.025768 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.025809 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.026280 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.026402 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.043497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl4c4\" (UniqueName: \"kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4\") pod \"redhat-operators-z5v8p\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.134741 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.578150 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.796434 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerStarted","Data":"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc"} Dec 12 07:55:07 crc kubenswrapper[4867]: I1212 07:55:07.796493 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerStarted","Data":"da0051fa317528a73ccb02ca965b8c2514a83ca959df00ad80e3614e4c9cae04"} Dec 12 07:55:08 crc kubenswrapper[4867]: I1212 07:55:08.807628 4867 generic.go:334] "Generic (PLEG): container finished" podID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerID="8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc" exitCode=0 Dec 12 07:55:08 crc kubenswrapper[4867]: I1212 07:55:08.807674 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerDied","Data":"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc"} Dec 12 07:55:09 crc kubenswrapper[4867]: I1212 07:55:09.815776 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerStarted","Data":"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f"} Dec 12 07:55:10 crc kubenswrapper[4867]: I1212 07:55:10.837756 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:55:10 crc kubenswrapper[4867]: E1212 07:55:10.837970 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:55:10 crc kubenswrapper[4867]: I1212 07:55:10.845039 4867 generic.go:334] "Generic (PLEG): container finished" podID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerID="1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f" exitCode=0 Dec 12 07:55:10 crc kubenswrapper[4867]: I1212 07:55:10.848634 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerDied","Data":"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f"} Dec 12 07:55:11 crc kubenswrapper[4867]: I1212 07:55:11.858076 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerStarted","Data":"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8"} Dec 12 07:55:11 crc kubenswrapper[4867]: I1212 07:55:11.876527 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z5v8p" podStartSLOduration=3.328932367 podStartE2EDuration="5.876508664s" podCreationTimestamp="2025-12-12 07:55:06 +0000 UTC" firstStartedPulling="2025-12-12 07:55:08.809758307 +0000 UTC m=+3996.381139576" lastFinishedPulling="2025-12-12 07:55:11.357334604 +0000 UTC m=+3998.928715873" observedRunningTime="2025-12-12 07:55:11.874188116 +0000 UTC m=+3999.445569405" watchObservedRunningTime="2025-12-12 07:55:11.876508664 +0000 UTC m=+3999.447889933" Dec 12 07:55:17 crc kubenswrapper[4867]: I1212 07:55:17.135560 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:17 crc kubenswrapper[4867]: I1212 07:55:17.136173 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:17 crc kubenswrapper[4867]: I1212 07:55:17.179908 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:17 crc kubenswrapper[4867]: I1212 07:55:17.935480 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:17 crc kubenswrapper[4867]: I1212 07:55:17.974484 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:19 crc kubenswrapper[4867]: I1212 07:55:19.911630 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z5v8p" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="registry-server" containerID="cri-o://db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8" gracePeriod=2 Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.295928 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.403873 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content\") pod \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.403927 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl4c4\" (UniqueName: \"kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4\") pod \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.404029 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities\") pod \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\" (UID: \"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59\") " Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.405369 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities" (OuterVolumeSpecName: "utilities") pod "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" (UID: "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.409465 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4" (OuterVolumeSpecName: "kube-api-access-xl4c4") pod "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" (UID: "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59"). InnerVolumeSpecName "kube-api-access-xl4c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.506191 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl4c4\" (UniqueName: \"kubernetes.io/projected/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-kube-api-access-xl4c4\") on node \"crc\" DevicePath \"\"" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.506591 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.919785 4867 generic.go:334] "Generic (PLEG): container finished" podID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerID="db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8" exitCode=0 Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.919835 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerDied","Data":"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8"} Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.919855 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z5v8p" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.919928 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z5v8p" event={"ID":"04c1f322-cf6c-492f-8b6b-c1a44a9d0b59","Type":"ContainerDied","Data":"da0051fa317528a73ccb02ca965b8c2514a83ca959df00ad80e3614e4c9cae04"} Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.919994 4867 scope.go:117] "RemoveContainer" containerID="db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.936504 4867 scope.go:117] "RemoveContainer" containerID="1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.954797 4867 scope.go:117] "RemoveContainer" containerID="8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.979772 4867 scope.go:117] "RemoveContainer" containerID="db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8" Dec 12 07:55:20 crc kubenswrapper[4867]: E1212 07:55:20.980242 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8\": container with ID starting with db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8 not found: ID does not exist" containerID="db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.980304 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8"} err="failed to get container status \"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8\": rpc error: code = NotFound desc = could not find container \"db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8\": container with ID starting with db3934dcf2290f4982a67138ba3cdc2518b486e7f18b6ad65f9e0018628e04f8 not found: ID does not exist" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.980331 4867 scope.go:117] "RemoveContainer" containerID="1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f" Dec 12 07:55:20 crc kubenswrapper[4867]: E1212 07:55:20.980764 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f\": container with ID starting with 1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f not found: ID does not exist" containerID="1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.980809 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f"} err="failed to get container status \"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f\": rpc error: code = NotFound desc = could not find container \"1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f\": container with ID starting with 1f16d936e8739b0662c382b198b41aa80142132a6f87f28cc35f8a73ac8a297f not found: ID does not exist" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.980836 4867 scope.go:117] "RemoveContainer" containerID="8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc" Dec 12 07:55:20 crc kubenswrapper[4867]: E1212 07:55:20.981112 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc\": container with ID starting with 8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc not found: ID does not exist" containerID="8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc" Dec 12 07:55:20 crc kubenswrapper[4867]: I1212 07:55:20.981166 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc"} err="failed to get container status \"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc\": rpc error: code = NotFound desc = could not find container \"8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc\": container with ID starting with 8c0e0f728260511825c2c240a8c1252c7173f6c297900a243b06885b5ff938dc not found: ID does not exist" Dec 12 07:55:21 crc kubenswrapper[4867]: I1212 07:55:21.526256 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" (UID: "04c1f322-cf6c-492f-8b6b-c1a44a9d0b59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 07:55:21 crc kubenswrapper[4867]: I1212 07:55:21.622363 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 07:55:21 crc kubenswrapper[4867]: I1212 07:55:21.850842 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:21 crc kubenswrapper[4867]: I1212 07:55:21.858757 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z5v8p"] Dec 12 07:55:22 crc kubenswrapper[4867]: I1212 07:55:22.845886 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:55:22 crc kubenswrapper[4867]: E1212 07:55:22.846521 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:55:22 crc kubenswrapper[4867]: I1212 07:55:22.852198 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" path="/var/lib/kubelet/pods/04c1f322-cf6c-492f-8b6b-c1a44a9d0b59/volumes" Dec 12 07:55:34 crc kubenswrapper[4867]: I1212 07:55:34.838497 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:55:34 crc kubenswrapper[4867]: E1212 07:55:34.839373 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:55:46 crc kubenswrapper[4867]: I1212 07:55:46.838181 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:55:46 crc kubenswrapper[4867]: E1212 07:55:46.838946 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:55:59 crc kubenswrapper[4867]: I1212 07:55:59.839058 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:55:59 crc kubenswrapper[4867]: E1212 07:55:59.839781 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:56:12 crc kubenswrapper[4867]: I1212 07:56:12.842005 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:56:12 crc kubenswrapper[4867]: E1212 07:56:12.842997 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:56:25 crc kubenswrapper[4867]: I1212 07:56:25.838898 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:56:25 crc kubenswrapper[4867]: E1212 07:56:25.839647 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:56:40 crc kubenswrapper[4867]: I1212 07:56:40.838423 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:56:40 crc kubenswrapper[4867]: E1212 07:56:40.839160 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:56:53 crc kubenswrapper[4867]: I1212 07:56:53.838301 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:56:53 crc kubenswrapper[4867]: E1212 07:56:53.838983 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:57:08 crc kubenswrapper[4867]: I1212 07:57:08.838914 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:57:08 crc kubenswrapper[4867]: E1212 07:57:08.841254 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:57:20 crc kubenswrapper[4867]: I1212 07:57:20.837736 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:57:20 crc kubenswrapper[4867]: E1212 07:57:20.838550 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 07:57:33 crc kubenswrapper[4867]: I1212 07:57:33.838213 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 07:57:34 crc kubenswrapper[4867]: I1212 07:57:34.849873 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed"} Dec 12 07:59:58 crc kubenswrapper[4867]: I1212 07:59:58.989025 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 07:59:58 crc kubenswrapper[4867]: I1212 07:59:58.989666 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.177556 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks"] Dec 12 08:00:00 crc kubenswrapper[4867]: E1212 08:00:00.178129 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="extract-content" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.178141 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="extract-content" Dec 12 08:00:00 crc kubenswrapper[4867]: E1212 08:00:00.178160 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="registry-server" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.178166 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="registry-server" Dec 12 08:00:00 crc kubenswrapper[4867]: E1212 08:00:00.178181 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="extract-utilities" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.178187 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="extract-utilities" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.178322 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="04c1f322-cf6c-492f-8b6b-c1a44a9d0b59" containerName="registry-server" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.178863 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.181447 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.182547 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.190787 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks"] Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.293179 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.293290 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.293351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5vcx\" (UniqueName: \"kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.394371 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5vcx\" (UniqueName: \"kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.394457 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.394491 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.395710 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.412193 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.416679 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5vcx\" (UniqueName: \"kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx\") pod \"collect-profiles-29425440-694ks\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.502184 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:00 crc kubenswrapper[4867]: I1212 08:00:00.909834 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks"] Dec 12 08:00:01 crc kubenswrapper[4867]: I1212 08:00:01.887626 4867 generic.go:334] "Generic (PLEG): container finished" podID="3487e86b-ff16-453e-8eb5-ce86d6491ab1" containerID="fff20affae3ac5e445bea8644333ece80900a81621c10fbb5152298c9b1b51cc" exitCode=0 Dec 12 08:00:01 crc kubenswrapper[4867]: I1212 08:00:01.887835 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" event={"ID":"3487e86b-ff16-453e-8eb5-ce86d6491ab1","Type":"ContainerDied","Data":"fff20affae3ac5e445bea8644333ece80900a81621c10fbb5152298c9b1b51cc"} Dec 12 08:00:01 crc kubenswrapper[4867]: I1212 08:00:01.887938 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" event={"ID":"3487e86b-ff16-453e-8eb5-ce86d6491ab1","Type":"ContainerStarted","Data":"cdff4d9725bddff6f2ee1cca65d3db7f188266fe5f5834d1363da0d2627fa30d"} Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.200763 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.329244 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5vcx\" (UniqueName: \"kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx\") pod \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.329331 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume\") pod \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.329364 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume\") pod \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\" (UID: \"3487e86b-ff16-453e-8eb5-ce86d6491ab1\") " Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.330190 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume" (OuterVolumeSpecName: "config-volume") pod "3487e86b-ff16-453e-8eb5-ce86d6491ab1" (UID: "3487e86b-ff16-453e-8eb5-ce86d6491ab1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.335351 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3487e86b-ff16-453e-8eb5-ce86d6491ab1" (UID: "3487e86b-ff16-453e-8eb5-ce86d6491ab1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.335429 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx" (OuterVolumeSpecName: "kube-api-access-j5vcx") pod "3487e86b-ff16-453e-8eb5-ce86d6491ab1" (UID: "3487e86b-ff16-453e-8eb5-ce86d6491ab1"). InnerVolumeSpecName "kube-api-access-j5vcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.430751 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5vcx\" (UniqueName: \"kubernetes.io/projected/3487e86b-ff16-453e-8eb5-ce86d6491ab1-kube-api-access-j5vcx\") on node \"crc\" DevicePath \"\"" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.430794 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3487e86b-ff16-453e-8eb5-ce86d6491ab1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.430806 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3487e86b-ff16-453e-8eb5-ce86d6491ab1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.903813 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" event={"ID":"3487e86b-ff16-453e-8eb5-ce86d6491ab1","Type":"ContainerDied","Data":"cdff4d9725bddff6f2ee1cca65d3db7f188266fe5f5834d1363da0d2627fa30d"} Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.904130 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdff4d9725bddff6f2ee1cca65d3db7f188266fe5f5834d1363da0d2627fa30d" Dec 12 08:00:03 crc kubenswrapper[4867]: I1212 08:00:03.903837 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks" Dec 12 08:00:04 crc kubenswrapper[4867]: I1212 08:00:04.286171 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59"] Dec 12 08:00:04 crc kubenswrapper[4867]: I1212 08:00:04.291435 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425395-qbb59"] Dec 12 08:00:04 crc kubenswrapper[4867]: I1212 08:00:04.860440 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9f4224a-94dd-44b0-a236-d329322edbf2" path="/var/lib/kubelet/pods/d9f4224a-94dd-44b0-a236-d329322edbf2/volumes" Dec 12 08:00:28 crc kubenswrapper[4867]: I1212 08:00:28.989222 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:00:28 crc kubenswrapper[4867]: I1212 08:00:28.989797 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:00:43 crc kubenswrapper[4867]: I1212 08:00:43.397126 4867 scope.go:117] "RemoveContainer" containerID="263906e1ec399971bc56c4340068b099c334fea081d41bee2129d19a6de510c4" Dec 12 08:00:58 crc kubenswrapper[4867]: I1212 08:00:58.988854 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:00:58 crc kubenswrapper[4867]: I1212 08:00:58.989412 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:00:58 crc kubenswrapper[4867]: I1212 08:00:58.989458 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:00:58 crc kubenswrapper[4867]: I1212 08:00:58.990024 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:00:58 crc kubenswrapper[4867]: I1212 08:00:58.990086 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed" gracePeriod=600 Dec 12 08:00:59 crc kubenswrapper[4867]: I1212 08:00:59.302751 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed" exitCode=0 Dec 12 08:00:59 crc kubenswrapper[4867]: I1212 08:00:59.302823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed"} Dec 12 08:00:59 crc kubenswrapper[4867]: I1212 08:00:59.303125 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f"} Dec 12 08:00:59 crc kubenswrapper[4867]: I1212 08:00:59.303187 4867 scope.go:117] "RemoveContainer" containerID="fa9c1493d050f93cda274ebabdc8089086426d53ee1d2bf0e35e2aa451b73dfb" Dec 12 08:03:28 crc kubenswrapper[4867]: I1212 08:03:28.988757 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:03:28 crc kubenswrapper[4867]: I1212 08:03:28.989328 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.758587 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:03:46 crc kubenswrapper[4867]: E1212 08:03:46.759411 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3487e86b-ff16-453e-8eb5-ce86d6491ab1" containerName="collect-profiles" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.759427 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3487e86b-ff16-453e-8eb5-ce86d6491ab1" containerName="collect-profiles" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.759622 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3487e86b-ff16-453e-8eb5-ce86d6491ab1" containerName="collect-profiles" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.760879 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.784200 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.873967 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.874102 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g76sg\" (UniqueName: \"kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.874172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.975391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g76sg\" (UniqueName: \"kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.975461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.975544 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.976101 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:46 crc kubenswrapper[4867]: I1212 08:03:46.976191 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:47 crc kubenswrapper[4867]: I1212 08:03:47.000271 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g76sg\" (UniqueName: \"kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg\") pod \"community-operators-586x8\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:47 crc kubenswrapper[4867]: I1212 08:03:47.083182 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:47 crc kubenswrapper[4867]: I1212 08:03:47.628922 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:03:48 crc kubenswrapper[4867]: I1212 08:03:48.489466 4867 generic.go:334] "Generic (PLEG): container finished" podID="929b95d0-bc52-4ffe-983f-daca553fca55" containerID="63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b" exitCode=0 Dec 12 08:03:48 crc kubenswrapper[4867]: I1212 08:03:48.489518 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerDied","Data":"63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b"} Dec 12 08:03:48 crc kubenswrapper[4867]: I1212 08:03:48.489799 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerStarted","Data":"24dd6c9318e65aefbf83ec81a421a082233ab8130720a48f5631ea591df52558"} Dec 12 08:03:48 crc kubenswrapper[4867]: I1212 08:03:48.491989 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:03:49 crc kubenswrapper[4867]: I1212 08:03:49.497183 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerStarted","Data":"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3"} Dec 12 08:03:50 crc kubenswrapper[4867]: I1212 08:03:50.506607 4867 generic.go:334] "Generic (PLEG): container finished" podID="929b95d0-bc52-4ffe-983f-daca553fca55" containerID="02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3" exitCode=0 Dec 12 08:03:50 crc kubenswrapper[4867]: I1212 08:03:50.506732 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerDied","Data":"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3"} Dec 12 08:03:51 crc kubenswrapper[4867]: I1212 08:03:51.515907 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerStarted","Data":"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3"} Dec 12 08:03:51 crc kubenswrapper[4867]: I1212 08:03:51.552083 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-586x8" podStartSLOduration=2.86635232 podStartE2EDuration="5.552064913s" podCreationTimestamp="2025-12-12 08:03:46 +0000 UTC" firstStartedPulling="2025-12-12 08:03:48.491731927 +0000 UTC m=+4516.063113196" lastFinishedPulling="2025-12-12 08:03:51.17744452 +0000 UTC m=+4518.748825789" observedRunningTime="2025-12-12 08:03:51.549126321 +0000 UTC m=+4519.120507600" watchObservedRunningTime="2025-12-12 08:03:51.552064913 +0000 UTC m=+4519.123446182" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.150554 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.152743 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.159648 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.281409 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.281509 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg5n6\" (UniqueName: \"kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.281632 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.383359 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.383460 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.383499 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg5n6\" (UniqueName: \"kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.383957 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.384004 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.413126 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg5n6\" (UniqueName: \"kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6\") pod \"redhat-marketplace-8m7pk\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.471539 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:03:54 crc kubenswrapper[4867]: I1212 08:03:54.947019 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:03:54 crc kubenswrapper[4867]: W1212 08:03:54.951442 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod387e38ca_b70a_4eae_8d2e_ad7f66e2ba61.slice/crio-12e9eaac420f8711f848242a767f9c8d5cd31b4ff0e0f6771c0f72a3f883c346 WatchSource:0}: Error finding container 12e9eaac420f8711f848242a767f9c8d5cd31b4ff0e0f6771c0f72a3f883c346: Status 404 returned error can't find the container with id 12e9eaac420f8711f848242a767f9c8d5cd31b4ff0e0f6771c0f72a3f883c346 Dec 12 08:03:55 crc kubenswrapper[4867]: I1212 08:03:55.548700 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerStarted","Data":"12e9eaac420f8711f848242a767f9c8d5cd31b4ff0e0f6771c0f72a3f883c346"} Dec 12 08:03:56 crc kubenswrapper[4867]: I1212 08:03:56.555542 4867 generic.go:334] "Generic (PLEG): container finished" podID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerID="2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9" exitCode=0 Dec 12 08:03:56 crc kubenswrapper[4867]: I1212 08:03:56.555588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerDied","Data":"2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9"} Dec 12 08:03:57 crc kubenswrapper[4867]: I1212 08:03:57.084188 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:57 crc kubenswrapper[4867]: I1212 08:03:57.084574 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:57 crc kubenswrapper[4867]: I1212 08:03:57.133762 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:57 crc kubenswrapper[4867]: I1212 08:03:57.606242 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:03:58 crc kubenswrapper[4867]: I1212 08:03:58.335037 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:03:58 crc kubenswrapper[4867]: I1212 08:03:58.571720 4867 generic.go:334] "Generic (PLEG): container finished" podID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerID="11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932" exitCode=0 Dec 12 08:03:58 crc kubenswrapper[4867]: I1212 08:03:58.572042 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerDied","Data":"11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932"} Dec 12 08:03:58 crc kubenswrapper[4867]: I1212 08:03:58.989015 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:03:58 crc kubenswrapper[4867]: I1212 08:03:58.989073 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:03:59 crc kubenswrapper[4867]: I1212 08:03:59.580259 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerStarted","Data":"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538"} Dec 12 08:03:59 crc kubenswrapper[4867]: I1212 08:03:59.580582 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-586x8" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="registry-server" containerID="cri-o://b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3" gracePeriod=2 Dec 12 08:03:59 crc kubenswrapper[4867]: I1212 08:03:59.601066 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8m7pk" podStartSLOduration=2.925745974 podStartE2EDuration="5.601049592s" podCreationTimestamp="2025-12-12 08:03:54 +0000 UTC" firstStartedPulling="2025-12-12 08:03:56.55800824 +0000 UTC m=+4524.129389509" lastFinishedPulling="2025-12-12 08:03:59.233311858 +0000 UTC m=+4526.804693127" observedRunningTime="2025-12-12 08:03:59.600804705 +0000 UTC m=+4527.172185984" watchObservedRunningTime="2025-12-12 08:03:59.601049592 +0000 UTC m=+4527.172430861" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.043811 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.159047 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content\") pod \"929b95d0-bc52-4ffe-983f-daca553fca55\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.159122 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities\") pod \"929b95d0-bc52-4ffe-983f-daca553fca55\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.159170 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g76sg\" (UniqueName: \"kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg\") pod \"929b95d0-bc52-4ffe-983f-daca553fca55\" (UID: \"929b95d0-bc52-4ffe-983f-daca553fca55\") " Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.159970 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities" (OuterVolumeSpecName: "utilities") pod "929b95d0-bc52-4ffe-983f-daca553fca55" (UID: "929b95d0-bc52-4ffe-983f-daca553fca55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.168548 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg" (OuterVolumeSpecName: "kube-api-access-g76sg") pod "929b95d0-bc52-4ffe-983f-daca553fca55" (UID: "929b95d0-bc52-4ffe-983f-daca553fca55"). InnerVolumeSpecName "kube-api-access-g76sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.213882 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "929b95d0-bc52-4ffe-983f-daca553fca55" (UID: "929b95d0-bc52-4ffe-983f-daca553fca55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.260717 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g76sg\" (UniqueName: \"kubernetes.io/projected/929b95d0-bc52-4ffe-983f-daca553fca55-kube-api-access-g76sg\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.260753 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.260766 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/929b95d0-bc52-4ffe-983f-daca553fca55-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.590247 4867 generic.go:334] "Generic (PLEG): container finished" podID="929b95d0-bc52-4ffe-983f-daca553fca55" containerID="b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3" exitCode=0 Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.590327 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-586x8" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.590383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerDied","Data":"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3"} Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.590418 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-586x8" event={"ID":"929b95d0-bc52-4ffe-983f-daca553fca55","Type":"ContainerDied","Data":"24dd6c9318e65aefbf83ec81a421a082233ab8130720a48f5631ea591df52558"} Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.590436 4867 scope.go:117] "RemoveContainer" containerID="b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.610922 4867 scope.go:117] "RemoveContainer" containerID="02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.636250 4867 scope.go:117] "RemoveContainer" containerID="63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.658206 4867 scope.go:117] "RemoveContainer" containerID="b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3" Dec 12 08:04:00 crc kubenswrapper[4867]: E1212 08:04:00.659520 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3\": container with ID starting with b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3 not found: ID does not exist" containerID="b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.659570 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3"} err="failed to get container status \"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3\": rpc error: code = NotFound desc = could not find container \"b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3\": container with ID starting with b890b0d6d7ae08ecc7e3c4b6996a46538f0ad9c972075c845791ccb1cf0fb8e3 not found: ID does not exist" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.659597 4867 scope.go:117] "RemoveContainer" containerID="02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3" Dec 12 08:04:00 crc kubenswrapper[4867]: E1212 08:04:00.659993 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3\": container with ID starting with 02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3 not found: ID does not exist" containerID="02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.660026 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3"} err="failed to get container status \"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3\": rpc error: code = NotFound desc = could not find container \"02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3\": container with ID starting with 02ad1faef8e3a405aedf631d2fb0113921a9fecb2a155d4a8c738b38c916ebb3 not found: ID does not exist" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.660050 4867 scope.go:117] "RemoveContainer" containerID="63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b" Dec 12 08:04:00 crc kubenswrapper[4867]: E1212 08:04:00.660467 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b\": container with ID starting with 63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b not found: ID does not exist" containerID="63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.660506 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b"} err="failed to get container status \"63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b\": rpc error: code = NotFound desc = could not find container \"63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b\": container with ID starting with 63a56ece33c257cb94dbb715c8162201bccfb8644ffa544a595be7030efdb30b not found: ID does not exist" Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.699978 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.712219 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-586x8"] Dec 12 08:04:00 crc kubenswrapper[4867]: I1212 08:04:00.848312 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" path="/var/lib/kubelet/pods/929b95d0-bc52-4ffe-983f-daca553fca55/volumes" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.707358 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-kdqtv"] Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.712108 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-kdqtv"] Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.837116 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-g67z7"] Dec 12 08:04:03 crc kubenswrapper[4867]: E1212 08:04:03.837472 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="extract-content" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.837490 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="extract-content" Dec 12 08:04:03 crc kubenswrapper[4867]: E1212 08:04:03.837516 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="extract-utilities" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.837523 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="extract-utilities" Dec 12 08:04:03 crc kubenswrapper[4867]: E1212 08:04:03.837538 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="registry-server" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.837543 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="registry-server" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.837674 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="929b95d0-bc52-4ffe-983f-daca553fca55" containerName="registry-server" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.838245 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.840212 4867 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-sdnsw" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.840261 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.840437 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.842602 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.851993 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g67z7"] Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.907417 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.907949 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:03 crc kubenswrapper[4867]: I1212 08:04:03.908548 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4tc4\" (UniqueName: \"kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.009976 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4tc4\" (UniqueName: \"kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.010080 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.010121 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.010449 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.011009 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.030029 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4tc4\" (UniqueName: \"kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4\") pod \"crc-storage-crc-g67z7\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.157076 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.471796 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.472087 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.547864 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.570507 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g67z7"] Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.618477 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g67z7" event={"ID":"6253469d-de96-4f0c-a362-fb4ebf6ed080","Type":"ContainerStarted","Data":"c961eec4e0c51cabf9227f1f80fb532c0fe4631459108e25555e9459848213fa"} Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.658820 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:04 crc kubenswrapper[4867]: I1212 08:04:04.852023 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0587769e-ee2a-40bd-8f22-0d8fbf2f883a" path="/var/lib/kubelet/pods/0587769e-ee2a-40bd-8f22-0d8fbf2f883a/volumes" Dec 12 08:04:05 crc kubenswrapper[4867]: I1212 08:04:05.736534 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:04:06 crc kubenswrapper[4867]: I1212 08:04:06.641214 4867 generic.go:334] "Generic (PLEG): container finished" podID="6253469d-de96-4f0c-a362-fb4ebf6ed080" containerID="34a3854595e4ea075be3c5425a69508087da5185d978234c76d048c4a392abcb" exitCode=0 Dec 12 08:04:06 crc kubenswrapper[4867]: I1212 08:04:06.641269 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g67z7" event={"ID":"6253469d-de96-4f0c-a362-fb4ebf6ed080","Type":"ContainerDied","Data":"34a3854595e4ea075be3c5425a69508087da5185d978234c76d048c4a392abcb"} Dec 12 08:04:06 crc kubenswrapper[4867]: I1212 08:04:06.641699 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8m7pk" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="registry-server" containerID="cri-o://71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538" gracePeriod=2 Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.080059 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.149989 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5n6\" (UniqueName: \"kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6\") pod \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.150066 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content\") pod \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.150177 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities\") pod \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\" (UID: \"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61\") " Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.155113 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6" (OuterVolumeSpecName: "kube-api-access-mg5n6") pod "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" (UID: "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61"). InnerVolumeSpecName "kube-api-access-mg5n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.163573 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities" (OuterVolumeSpecName: "utilities") pod "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" (UID: "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.180846 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" (UID: "387e38ca-b70a-4eae-8d2e-ad7f66e2ba61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.251706 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.251740 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5n6\" (UniqueName: \"kubernetes.io/projected/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-kube-api-access-mg5n6\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.251749 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.648299 4867 generic.go:334] "Generic (PLEG): container finished" podID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerID="71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538" exitCode=0 Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.648390 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8m7pk" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.648421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerDied","Data":"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538"} Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.648481 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8m7pk" event={"ID":"387e38ca-b70a-4eae-8d2e-ad7f66e2ba61","Type":"ContainerDied","Data":"12e9eaac420f8711f848242a767f9c8d5cd31b4ff0e0f6771c0f72a3f883c346"} Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.648503 4867 scope.go:117] "RemoveContainer" containerID="71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.665847 4867 scope.go:117] "RemoveContainer" containerID="11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.682490 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.690922 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8m7pk"] Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.705305 4867 scope.go:117] "RemoveContainer" containerID="2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.776022 4867 scope.go:117] "RemoveContainer" containerID="71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538" Dec 12 08:04:07 crc kubenswrapper[4867]: E1212 08:04:07.776510 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538\": container with ID starting with 71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538 not found: ID does not exist" containerID="71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.776551 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538"} err="failed to get container status \"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538\": rpc error: code = NotFound desc = could not find container \"71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538\": container with ID starting with 71adffad8ece37784a9a0d988f541f275b03358f625e5d083b02152d4b28a538 not found: ID does not exist" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.776578 4867 scope.go:117] "RemoveContainer" containerID="11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932" Dec 12 08:04:07 crc kubenswrapper[4867]: E1212 08:04:07.777504 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932\": container with ID starting with 11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932 not found: ID does not exist" containerID="11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.777529 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932"} err="failed to get container status \"11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932\": rpc error: code = NotFound desc = could not find container \"11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932\": container with ID starting with 11cb08c6fd40af0843e584eb233ba7863797c8689736b8becf48e538d9bca932 not found: ID does not exist" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.777547 4867 scope.go:117] "RemoveContainer" containerID="2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9" Dec 12 08:04:07 crc kubenswrapper[4867]: E1212 08:04:07.777847 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9\": container with ID starting with 2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9 not found: ID does not exist" containerID="2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.777893 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9"} err="failed to get container status \"2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9\": rpc error: code = NotFound desc = could not find container \"2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9\": container with ID starting with 2eec7904e946dc92ac5a175e53fb53a6bc7b681ee46e4452b48819892a5d2db9 not found: ID does not exist" Dec 12 08:04:07 crc kubenswrapper[4867]: I1212 08:04:07.977500 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.168404 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt\") pod \"6253469d-de96-4f0c-a362-fb4ebf6ed080\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.168577 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage\") pod \"6253469d-de96-4f0c-a362-fb4ebf6ed080\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.168681 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4tc4\" (UniqueName: \"kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4\") pod \"6253469d-de96-4f0c-a362-fb4ebf6ed080\" (UID: \"6253469d-de96-4f0c-a362-fb4ebf6ed080\") " Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.168798 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "6253469d-de96-4f0c-a362-fb4ebf6ed080" (UID: "6253469d-de96-4f0c-a362-fb4ebf6ed080"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.169086 4867 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/6253469d-de96-4f0c-a362-fb4ebf6ed080-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.176084 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4" (OuterVolumeSpecName: "kube-api-access-c4tc4") pod "6253469d-de96-4f0c-a362-fb4ebf6ed080" (UID: "6253469d-de96-4f0c-a362-fb4ebf6ed080"). InnerVolumeSpecName "kube-api-access-c4tc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.190739 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "6253469d-de96-4f0c-a362-fb4ebf6ed080" (UID: "6253469d-de96-4f0c-a362-fb4ebf6ed080"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.270690 4867 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/6253469d-de96-4f0c-a362-fb4ebf6ed080-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.270738 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4tc4\" (UniqueName: \"kubernetes.io/projected/6253469d-de96-4f0c-a362-fb4ebf6ed080-kube-api-access-c4tc4\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.656878 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g67z7" event={"ID":"6253469d-de96-4f0c-a362-fb4ebf6ed080","Type":"ContainerDied","Data":"c961eec4e0c51cabf9227f1f80fb532c0fe4631459108e25555e9459848213fa"} Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.657174 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c961eec4e0c51cabf9227f1f80fb532c0fe4631459108e25555e9459848213fa" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.656929 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g67z7" Dec 12 08:04:08 crc kubenswrapper[4867]: I1212 08:04:08.847831 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" path="/var/lib/kubelet/pods/387e38ca-b70a-4eae-8d2e-ad7f66e2ba61/volumes" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.105726 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-g67z7"] Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.111974 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-g67z7"] Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147062 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:10 crc kubenswrapper[4867]: E1212 08:04:10.147434 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="extract-content" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147449 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="extract-content" Dec 12 08:04:10 crc kubenswrapper[4867]: E1212 08:04:10.147464 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="registry-server" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147472 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="registry-server" Dec 12 08:04:10 crc kubenswrapper[4867]: E1212 08:04:10.147482 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6253469d-de96-4f0c-a362-fb4ebf6ed080" containerName="storage" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147490 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6253469d-de96-4f0c-a362-fb4ebf6ed080" containerName="storage" Dec 12 08:04:10 crc kubenswrapper[4867]: E1212 08:04:10.147508 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="extract-utilities" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147517 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="extract-utilities" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147709 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="387e38ca-b70a-4eae-8d2e-ad7f66e2ba61" containerName="registry-server" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.147729 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6253469d-de96-4f0c-a362-fb4ebf6ed080" containerName="storage" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.148954 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.211680 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.266811 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-9ld7r"] Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.267702 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.269730 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.272410 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.273625 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.273670 4867 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-sdnsw" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.275982 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9ld7r"] Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.306496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.306571 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.306607 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkg2r\" (UniqueName: \"kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408129 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408194 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408239 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408262 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkg2r\" (UniqueName: \"kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408283 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408333 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88k56\" (UniqueName: \"kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408732 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.408890 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.453214 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkg2r\" (UniqueName: \"kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r\") pod \"certified-operators-68jlp\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.509594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88k56\" (UniqueName: \"kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.509978 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.510004 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.510306 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.510690 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.512190 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.539891 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88k56\" (UniqueName: \"kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56\") pod \"crc-storage-crc-9ld7r\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.581517 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.850982 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6253469d-de96-4f0c-a362-fb4ebf6ed080" path="/var/lib/kubelet/pods/6253469d-de96-4f0c-a362-fb4ebf6ed080/volumes" Dec 12 08:04:10 crc kubenswrapper[4867]: I1212 08:04:10.971536 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:11 crc kubenswrapper[4867]: I1212 08:04:11.070174 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9ld7r"] Dec 12 08:04:11 crc kubenswrapper[4867]: W1212 08:04:11.073118 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c141e6d_5082_4bc0_a377_9e2e32d441ed.slice/crio-f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8 WatchSource:0}: Error finding container f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8: Status 404 returned error can't find the container with id f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8 Dec 12 08:04:11 crc kubenswrapper[4867]: I1212 08:04:11.675492 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9ld7r" event={"ID":"9c141e6d-5082-4bc0-a377-9e2e32d441ed","Type":"ContainerStarted","Data":"f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8"} Dec 12 08:04:11 crc kubenswrapper[4867]: I1212 08:04:11.676858 4867 generic.go:334] "Generic (PLEG): container finished" podID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerID="4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862" exitCode=0 Dec 12 08:04:11 crc kubenswrapper[4867]: I1212 08:04:11.676909 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerDied","Data":"4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862"} Dec 12 08:04:11 crc kubenswrapper[4867]: I1212 08:04:11.676941 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerStarted","Data":"2cc76552f6c5ced75db3e053d4705cd2a862b6668d243a569c7897ab4653b974"} Dec 12 08:04:12 crc kubenswrapper[4867]: I1212 08:04:12.684884 4867 generic.go:334] "Generic (PLEG): container finished" podID="9c141e6d-5082-4bc0-a377-9e2e32d441ed" containerID="52f19d9e087e1327a65549d69c10044e4e8a617afdd30cb4e8b3ffa60963743a" exitCode=0 Dec 12 08:04:12 crc kubenswrapper[4867]: I1212 08:04:12.684995 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9ld7r" event={"ID":"9c141e6d-5082-4bc0-a377-9e2e32d441ed","Type":"ContainerDied","Data":"52f19d9e087e1327a65549d69c10044e4e8a617afdd30cb4e8b3ffa60963743a"} Dec 12 08:04:13 crc kubenswrapper[4867]: I1212 08:04:13.693589 4867 generic.go:334] "Generic (PLEG): container finished" podID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerID="2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2" exitCode=0 Dec 12 08:04:13 crc kubenswrapper[4867]: I1212 08:04:13.693681 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerDied","Data":"2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2"} Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.298986 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.368161 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage\") pod \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.368204 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88k56\" (UniqueName: \"kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56\") pod \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.368252 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt\") pod \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\" (UID: \"9c141e6d-5082-4bc0-a377-9e2e32d441ed\") " Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.368478 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9c141e6d-5082-4bc0-a377-9e2e32d441ed" (UID: "9c141e6d-5082-4bc0-a377-9e2e32d441ed"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.373478 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56" (OuterVolumeSpecName: "kube-api-access-88k56") pod "9c141e6d-5082-4bc0-a377-9e2e32d441ed" (UID: "9c141e6d-5082-4bc0-a377-9e2e32d441ed"). InnerVolumeSpecName "kube-api-access-88k56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.387803 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9c141e6d-5082-4bc0-a377-9e2e32d441ed" (UID: "9c141e6d-5082-4bc0-a377-9e2e32d441ed"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.469111 4867 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9c141e6d-5082-4bc0-a377-9e2e32d441ed-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.469144 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88k56\" (UniqueName: \"kubernetes.io/projected/9c141e6d-5082-4bc0-a377-9e2e32d441ed-kube-api-access-88k56\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.469155 4867 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9c141e6d-5082-4bc0-a377-9e2e32d441ed-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.701986 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9ld7r" event={"ID":"9c141e6d-5082-4bc0-a377-9e2e32d441ed","Type":"ContainerDied","Data":"f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8"} Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.702059 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f19218bf0bf49e936676c4777ba604270663a24f589deee544e9ecff2c640cf8" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.702003 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9ld7r" Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.703973 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerStarted","Data":"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede"} Dec 12 08:04:14 crc kubenswrapper[4867]: I1212 08:04:14.729423 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-68jlp" podStartSLOduration=2.017593585 podStartE2EDuration="4.729404099s" podCreationTimestamp="2025-12-12 08:04:10 +0000 UTC" firstStartedPulling="2025-12-12 08:04:11.678421564 +0000 UTC m=+4539.249802833" lastFinishedPulling="2025-12-12 08:04:14.390232078 +0000 UTC m=+4541.961613347" observedRunningTime="2025-12-12 08:04:14.724891579 +0000 UTC m=+4542.296272858" watchObservedRunningTime="2025-12-12 08:04:14.729404099 +0000 UTC m=+4542.300785368" Dec 12 08:04:20 crc kubenswrapper[4867]: I1212 08:04:20.513012 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:20 crc kubenswrapper[4867]: I1212 08:04:20.513432 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:20 crc kubenswrapper[4867]: I1212 08:04:20.554034 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:20 crc kubenswrapper[4867]: I1212 08:04:20.780234 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:20 crc kubenswrapper[4867]: I1212 08:04:20.836211 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:22 crc kubenswrapper[4867]: I1212 08:04:22.750010 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-68jlp" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="registry-server" containerID="cri-o://51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede" gracePeriod=2 Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.158019 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.292578 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content\") pod \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.292651 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities\") pod \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.292805 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkg2r\" (UniqueName: \"kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r\") pod \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\" (UID: \"ae1ddf86-8f0d-4844-9da9-3d3844981c92\") " Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.293722 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities" (OuterVolumeSpecName: "utilities") pod "ae1ddf86-8f0d-4844-9da9-3d3844981c92" (UID: "ae1ddf86-8f0d-4844-9da9-3d3844981c92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.312360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r" (OuterVolumeSpecName: "kube-api-access-fkg2r") pod "ae1ddf86-8f0d-4844-9da9-3d3844981c92" (UID: "ae1ddf86-8f0d-4844-9da9-3d3844981c92"). InnerVolumeSpecName "kube-api-access-fkg2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.356301 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae1ddf86-8f0d-4844-9da9-3d3844981c92" (UID: "ae1ddf86-8f0d-4844-9da9-3d3844981c92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.394161 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkg2r\" (UniqueName: \"kubernetes.io/projected/ae1ddf86-8f0d-4844-9da9-3d3844981c92-kube-api-access-fkg2r\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.394196 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.394205 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae1ddf86-8f0d-4844-9da9-3d3844981c92-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.764628 4867 generic.go:334] "Generic (PLEG): container finished" podID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerID="51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede" exitCode=0 Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.764695 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerDied","Data":"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede"} Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.764727 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68jlp" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.764752 4867 scope.go:117] "RemoveContainer" containerID="51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.764736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68jlp" event={"ID":"ae1ddf86-8f0d-4844-9da9-3d3844981c92","Type":"ContainerDied","Data":"2cc76552f6c5ced75db3e053d4705cd2a862b6668d243a569c7897ab4653b974"} Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.786220 4867 scope.go:117] "RemoveContainer" containerID="2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.805090 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.813346 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-68jlp"] Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.833947 4867 scope.go:117] "RemoveContainer" containerID="4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.850757 4867 scope.go:117] "RemoveContainer" containerID="51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede" Dec 12 08:04:23 crc kubenswrapper[4867]: E1212 08:04:23.851287 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede\": container with ID starting with 51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede not found: ID does not exist" containerID="51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.851321 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede"} err="failed to get container status \"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede\": rpc error: code = NotFound desc = could not find container \"51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede\": container with ID starting with 51a85200f31c32a16a6c28c1906d78fc460b0d2cf4298c21bc28462e6431dede not found: ID does not exist" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.851342 4867 scope.go:117] "RemoveContainer" containerID="2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2" Dec 12 08:04:23 crc kubenswrapper[4867]: E1212 08:04:23.851556 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2\": container with ID starting with 2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2 not found: ID does not exist" containerID="2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.851583 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2"} err="failed to get container status \"2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2\": rpc error: code = NotFound desc = could not find container \"2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2\": container with ID starting with 2776e651426dd10fd7f58cc20518190a109599ee254f9b03d0dcc0f995d4acc2 not found: ID does not exist" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.851603 4867 scope.go:117] "RemoveContainer" containerID="4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862" Dec 12 08:04:23 crc kubenswrapper[4867]: E1212 08:04:23.852545 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862\": container with ID starting with 4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862 not found: ID does not exist" containerID="4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862" Dec 12 08:04:23 crc kubenswrapper[4867]: I1212 08:04:23.852575 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862"} err="failed to get container status \"4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862\": rpc error: code = NotFound desc = could not find container \"4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862\": container with ID starting with 4a00672f6d4ba3e7974c524261257e708f52a4ea3f83aec1733e40f1ba717862 not found: ID does not exist" Dec 12 08:04:24 crc kubenswrapper[4867]: I1212 08:04:24.846912 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" path="/var/lib/kubelet/pods/ae1ddf86-8f0d-4844-9da9-3d3844981c92/volumes" Dec 12 08:04:28 crc kubenswrapper[4867]: I1212 08:04:28.988661 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:04:28 crc kubenswrapper[4867]: I1212 08:04:28.988991 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:04:28 crc kubenswrapper[4867]: I1212 08:04:28.989034 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:04:28 crc kubenswrapper[4867]: I1212 08:04:28.989683 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:04:28 crc kubenswrapper[4867]: I1212 08:04:28.989736 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" gracePeriod=600 Dec 12 08:04:29 crc kubenswrapper[4867]: E1212 08:04:29.111252 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:04:29 crc kubenswrapper[4867]: I1212 08:04:29.812842 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" exitCode=0 Dec 12 08:04:29 crc kubenswrapper[4867]: I1212 08:04:29.812937 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f"} Dec 12 08:04:29 crc kubenswrapper[4867]: I1212 08:04:29.813314 4867 scope.go:117] "RemoveContainer" containerID="1099c363960d5bb00d46e2bfab95f14538b9057aa3213f8f3ee5d9bb7f1b4aed" Dec 12 08:04:29 crc kubenswrapper[4867]: I1212 08:04:29.813821 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:04:29 crc kubenswrapper[4867]: E1212 08:04:29.814024 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:04:43 crc kubenswrapper[4867]: I1212 08:04:43.497721 4867 scope.go:117] "RemoveContainer" containerID="5e2bb9506306aa9fc906723b6edb756c61d2ea95e9ff94842567322e6385105d" Dec 12 08:04:43 crc kubenswrapper[4867]: I1212 08:04:43.837902 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:04:43 crc kubenswrapper[4867]: E1212 08:04:43.838433 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:04:55 crc kubenswrapper[4867]: I1212 08:04:55.838364 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:04:55 crc kubenswrapper[4867]: E1212 08:04:55.839341 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:05:10 crc kubenswrapper[4867]: I1212 08:05:10.839959 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:05:10 crc kubenswrapper[4867]: E1212 08:05:10.841352 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:05:25 crc kubenswrapper[4867]: I1212 08:05:25.839291 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:05:25 crc kubenswrapper[4867]: E1212 08:05:25.840587 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:05:36 crc kubenswrapper[4867]: I1212 08:05:36.838513 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:05:36 crc kubenswrapper[4867]: E1212 08:05:36.839214 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:05:47 crc kubenswrapper[4867]: I1212 08:05:47.838922 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:05:47 crc kubenswrapper[4867]: E1212 08:05:47.839634 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.934465 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:05:54 crc kubenswrapper[4867]: E1212 08:05:54.935405 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="extract-content" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935423 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="extract-content" Dec 12 08:05:54 crc kubenswrapper[4867]: E1212 08:05:54.935461 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="extract-utilities" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935469 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="extract-utilities" Dec 12 08:05:54 crc kubenswrapper[4867]: E1212 08:05:54.935486 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c141e6d-5082-4bc0-a377-9e2e32d441ed" containerName="storage" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935494 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c141e6d-5082-4bc0-a377-9e2e32d441ed" containerName="storage" Dec 12 08:05:54 crc kubenswrapper[4867]: E1212 08:05:54.935512 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="registry-server" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935519 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="registry-server" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935685 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c141e6d-5082-4bc0-a377-9e2e32d441ed" containerName="storage" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.935703 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1ddf86-8f0d-4844-9da9-3d3844981c92" containerName="registry-server" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.936960 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.961056 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.984970 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.985111 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:54 crc kubenswrapper[4867]: I1212 08:05:54.985327 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cblcr\" (UniqueName: \"kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.086796 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cblcr\" (UniqueName: \"kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.087051 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.087086 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.087582 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.087699 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.107432 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cblcr\" (UniqueName: \"kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr\") pod \"redhat-operators-v765s\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.253906 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:05:55 crc kubenswrapper[4867]: I1212 08:05:55.703876 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:05:56 crc kubenswrapper[4867]: I1212 08:05:56.449555 4867 generic.go:334] "Generic (PLEG): container finished" podID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerID="9798e886b15e2916b92b1184f489727f58872cf9f38c0179ebeb168393623caa" exitCode=0 Dec 12 08:05:56 crc kubenswrapper[4867]: I1212 08:05:56.449616 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerDied","Data":"9798e886b15e2916b92b1184f489727f58872cf9f38c0179ebeb168393623caa"} Dec 12 08:05:56 crc kubenswrapper[4867]: I1212 08:05:56.449653 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerStarted","Data":"b1c2d5f16ee8958c803b5e8e58939fdf689610b6b073de6be4d0177c4a688ecb"} Dec 12 08:05:57 crc kubenswrapper[4867]: I1212 08:05:57.457871 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerStarted","Data":"b3e5fa1d8b444763c74c9958142058225bd6d03e6ada12fcb2501e613f78e8b3"} Dec 12 08:05:58 crc kubenswrapper[4867]: I1212 08:05:58.466577 4867 generic.go:334] "Generic (PLEG): container finished" podID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerID="b3e5fa1d8b444763c74c9958142058225bd6d03e6ada12fcb2501e613f78e8b3" exitCode=0 Dec 12 08:05:58 crc kubenswrapper[4867]: I1212 08:05:58.466613 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerDied","Data":"b3e5fa1d8b444763c74c9958142058225bd6d03e6ada12fcb2501e613f78e8b3"} Dec 12 08:05:59 crc kubenswrapper[4867]: I1212 08:05:59.475555 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerStarted","Data":"6fe3cc0c93c337e9b581b91ddcc2d58428be208d3e9116fa0b8289d6c332d326"} Dec 12 08:05:59 crc kubenswrapper[4867]: I1212 08:05:59.491446 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v765s" podStartSLOduration=2.996106471 podStartE2EDuration="5.491430277s" podCreationTimestamp="2025-12-12 08:05:54 +0000 UTC" firstStartedPulling="2025-12-12 08:05:56.452026786 +0000 UTC m=+4644.023408055" lastFinishedPulling="2025-12-12 08:05:58.947350592 +0000 UTC m=+4646.518731861" observedRunningTime="2025-12-12 08:05:59.489881629 +0000 UTC m=+4647.061262918" watchObservedRunningTime="2025-12-12 08:05:59.491430277 +0000 UTC m=+4647.062811546" Dec 12 08:06:02 crc kubenswrapper[4867]: I1212 08:06:02.842108 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:06:02 crc kubenswrapper[4867]: E1212 08:06:02.842694 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:06:05 crc kubenswrapper[4867]: I1212 08:06:05.254323 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:05 crc kubenswrapper[4867]: I1212 08:06:05.254658 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:05 crc kubenswrapper[4867]: I1212 08:06:05.292635 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:05 crc kubenswrapper[4867]: I1212 08:06:05.560075 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:05 crc kubenswrapper[4867]: I1212 08:06:05.604917 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:06:07 crc kubenswrapper[4867]: I1212 08:06:07.539297 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v765s" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="registry-server" containerID="cri-o://6fe3cc0c93c337e9b581b91ddcc2d58428be208d3e9116fa0b8289d6c332d326" gracePeriod=2 Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.569112 4867 generic.go:334] "Generic (PLEG): container finished" podID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerID="6fe3cc0c93c337e9b581b91ddcc2d58428be208d3e9116fa0b8289d6c332d326" exitCode=0 Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.569195 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerDied","Data":"6fe3cc0c93c337e9b581b91ddcc2d58428be208d3e9116fa0b8289d6c332d326"} Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.665749 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.711105 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cblcr\" (UniqueName: \"kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr\") pod \"4bc55757-1611-48be-bb1a-3e43fbc07086\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.711216 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities\") pod \"4bc55757-1611-48be-bb1a-3e43fbc07086\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.711282 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content\") pod \"4bc55757-1611-48be-bb1a-3e43fbc07086\" (UID: \"4bc55757-1611-48be-bb1a-3e43fbc07086\") " Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.712152 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities" (OuterVolumeSpecName: "utilities") pod "4bc55757-1611-48be-bb1a-3e43fbc07086" (UID: "4bc55757-1611-48be-bb1a-3e43fbc07086"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.716426 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr" (OuterVolumeSpecName: "kube-api-access-cblcr") pod "4bc55757-1611-48be-bb1a-3e43fbc07086" (UID: "4bc55757-1611-48be-bb1a-3e43fbc07086"). InnerVolumeSpecName "kube-api-access-cblcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.813598 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cblcr\" (UniqueName: \"kubernetes.io/projected/4bc55757-1611-48be-bb1a-3e43fbc07086-kube-api-access-cblcr\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.813638 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.834658 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bc55757-1611-48be-bb1a-3e43fbc07086" (UID: "4bc55757-1611-48be-bb1a-3e43fbc07086"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:06:10 crc kubenswrapper[4867]: I1212 08:06:10.916029 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bc55757-1611-48be-bb1a-3e43fbc07086-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.577951 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v765s" event={"ID":"4bc55757-1611-48be-bb1a-3e43fbc07086","Type":"ContainerDied","Data":"b1c2d5f16ee8958c803b5e8e58939fdf689610b6b073de6be4d0177c4a688ecb"} Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.578002 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v765s" Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.578018 4867 scope.go:117] "RemoveContainer" containerID="6fe3cc0c93c337e9b581b91ddcc2d58428be208d3e9116fa0b8289d6c332d326" Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.602641 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.606755 4867 scope.go:117] "RemoveContainer" containerID="b3e5fa1d8b444763c74c9958142058225bd6d03e6ada12fcb2501e613f78e8b3" Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.608475 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v765s"] Dec 12 08:06:11 crc kubenswrapper[4867]: I1212 08:06:11.629675 4867 scope.go:117] "RemoveContainer" containerID="9798e886b15e2916b92b1184f489727f58872cf9f38c0179ebeb168393623caa" Dec 12 08:06:12 crc kubenswrapper[4867]: I1212 08:06:12.847269 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" path="/var/lib/kubelet/pods/4bc55757-1611-48be-bb1a-3e43fbc07086/volumes" Dec 12 08:06:13 crc kubenswrapper[4867]: I1212 08:06:13.838069 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:06:13 crc kubenswrapper[4867]: E1212 08:06:13.838366 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.004926 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:17 crc kubenswrapper[4867]: E1212 08:06:17.007860 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="extract-content" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.007884 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="extract-content" Dec 12 08:06:17 crc kubenswrapper[4867]: E1212 08:06:17.007899 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="registry-server" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.007906 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="registry-server" Dec 12 08:06:17 crc kubenswrapper[4867]: E1212 08:06:17.007921 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="extract-utilities" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.007932 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="extract-utilities" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.008093 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc55757-1611-48be-bb1a-3e43fbc07086" containerName="registry-server" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.008840 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.011654 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-9b2rw" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.011870 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.012109 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.012281 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.014368 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.015844 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.020816 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.021052 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.036874 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.110581 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4npm\" (UniqueName: \"kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.110687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.110733 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.110777 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.110802 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbjpj\" (UniqueName: \"kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.214918 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.214979 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbjpj\" (UniqueName: \"kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.215018 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4npm\" (UniqueName: \"kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.215064 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.215095 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.215893 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.216704 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.217402 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.248312 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbjpj\" (UniqueName: \"kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj\") pod \"dnsmasq-dns-7c95485ddf-lq2xg\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.249376 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4npm\" (UniqueName: \"kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm\") pod \"dnsmasq-dns-6f6d89767f-2zpft\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.283111 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.283609 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.302202 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.303485 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.364066 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.372205 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.418197 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpp9s\" (UniqueName: \"kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.418299 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.418354 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.519139 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpp9s\" (UniqueName: \"kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.519603 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.519633 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.520544 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.520650 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.539326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpp9s\" (UniqueName: \"kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s\") pod \"dnsmasq-dns-f5bc65899-mb2tr\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.680666 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.845665 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.851449 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.889419 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.903354 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.903953 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:06:17 crc kubenswrapper[4867]: I1212 08:06:17.927155 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.039083 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkbkr\" (UniqueName: \"kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.039172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.039241 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.141868 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.142518 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkbkr\" (UniqueName: \"kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.142659 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.142851 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.143400 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.160276 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkbkr\" (UniqueName: \"kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr\") pod \"dnsmasq-dns-55c47c755-pxw57\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.212901 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.224709 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.451572 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.453202 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.458706 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.458812 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.458954 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wklzn" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.459010 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.459460 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.459703 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.461885 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.466117 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.549951 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550017 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550048 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550086 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550106 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550128 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550171 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550194 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550217 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.550351 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hklsv\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: W1212 08:06:18.612708 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48f6e308_b8de_4401_817e_1b7201d6b13e.slice/crio-8087338e061c95283ce3c115f7734bd000372b55e2c61e9257ed881e0092a3f6 WatchSource:0}: Error finding container 8087338e061c95283ce3c115f7734bd000372b55e2c61e9257ed881e0092a3f6: Status 404 returned error can't find the container with id 8087338e061c95283ce3c115f7734bd000372b55e2c61e9257ed881e0092a3f6 Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651007 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" event={"ID":"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2","Type":"ContainerStarted","Data":"8204b65dd9ddbad2929d5bb89d19fb421c97a6425bc0e346686e70128b327cfd"} Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651380 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651497 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hklsv\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651532 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651563 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651589 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651633 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651654 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651679 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651718 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.651749 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.654047 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.654137 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.654598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.654758 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.654916 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.656860 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.656898 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ade0757b040b265f114a1895a4de61d5d63bae711e8c14acb63c5e5d4e3b9f57/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.658204 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.658949 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.659110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.660784 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.661297 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" event={"ID":"48f6e308-b8de-4401-817e-1b7201d6b13e","Type":"ContainerStarted","Data":"8087338e061c95283ce3c115f7734bd000372b55e2c61e9257ed881e0092a3f6"} Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.665491 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" event={"ID":"eb845851-49b6-490d-b58b-e8ae0419d94c","Type":"ContainerStarted","Data":"a724406f4c3d79392679fa9133a92e5dcb7ff276148d1fb5da7458285b3cc22d"} Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.670672 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hklsv\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.701075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.781751 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.987421 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.989079 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.991676 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.992547 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.992805 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.993569 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.993672 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.994359 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cjs79" Dec 12 08:06:18 crc kubenswrapper[4867]: I1212 08:06:18.998756 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.008326 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.057921 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.057962 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.057995 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bsjp\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058021 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058040 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058061 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058082 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058109 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058126 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058155 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.058173 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.159747 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bsjp\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160194 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160261 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160336 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160370 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160436 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160530 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160571 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160673 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.160725 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.162285 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.163176 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.163539 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.163777 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.168130 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.168171 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f952049225bdbbdac120d8064eb4b6b35c89d69f6e2090c2cba213d8afde8e90/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.172890 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.173179 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.173186 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.178280 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.180920 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bsjp\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.184904 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.216802 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.338256 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.442221 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.603655 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:06:19 crc kubenswrapper[4867]: W1212 08:06:19.614291 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2300b08d_c21e_4cf2_bca3_97440b179709.slice/crio-9e2c3a24f1d4ac6ee9852749a445a8d021dd4c9efd389e9d82e7310ec3a75f02 WatchSource:0}: Error finding container 9e2c3a24f1d4ac6ee9852749a445a8d021dd4c9efd389e9d82e7310ec3a75f02: Status 404 returned error can't find the container with id 9e2c3a24f1d4ac6ee9852749a445a8d021dd4c9efd389e9d82e7310ec3a75f02 Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.685912 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerStarted","Data":"9e2c3a24f1d4ac6ee9852749a445a8d021dd4c9efd389e9d82e7310ec3a75f02"} Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.690968 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c47c755-pxw57" event={"ID":"36db5392-1167-463c-ab34-f97bcc329454","Type":"ContainerStarted","Data":"0b7ebabfeda3511e0a73117b2c248842bb0d044d957aae5a9597bf35e92a1dcb"} Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.904366 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.936116 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.937614 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.942397 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lwwjr" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.942987 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.943997 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.944114 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.944196 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.948581 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975234 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4xdq\" (UniqueName: \"kubernetes.io/projected/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kube-api-access-p4xdq\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975309 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975346 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975381 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975422 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975458 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975728 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-default\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:19 crc kubenswrapper[4867]: I1212 08:06:19.975789 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kolla-config\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077002 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4xdq\" (UniqueName: \"kubernetes.io/projected/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kube-api-access-p4xdq\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077062 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077089 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077127 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077162 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077190 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077240 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-default\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077265 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kolla-config\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.077732 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.078304 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kolla-config\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.078343 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-config-data-default\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.081526 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.081557 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c08b01934e3712cf2ea78344d3c572465c54137547912c0f518150a6515647f0/globalmount\"" pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.086976 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.206801 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.212661 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4xdq\" (UniqueName: \"kubernetes.io/projected/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-kube-api-access-p4xdq\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.213587 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0583cfff-0ec3-4fad-84de-3c9f6f3dd49c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.241716 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fda8a5ac-1837-4823-b0c4-94511b67020a\") pod \"openstack-galera-0\" (UID: \"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c\") " pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.264589 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.710981 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerStarted","Data":"f7c31e3aa2319e09d3b415f6f2b367bedfa320d3674b4f923e6c1031f962386c"} Dec 12 08:06:20 crc kubenswrapper[4867]: I1212 08:06:20.853265 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.454175 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.457160 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.461942 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.462283 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-pw58c" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.462505 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.463669 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.466572 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601442 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601519 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601567 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601589 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601607 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2njjp\" (UniqueName: \"kubernetes.io/projected/07166244-e179-4d78-a254-942fcf7bee79-kube-api-access-2njjp\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601667 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601689 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.601729 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/07166244-e179-4d78-a254-942fcf7bee79-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703711 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703795 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703825 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703847 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2njjp\" (UniqueName: \"kubernetes.io/projected/07166244-e179-4d78-a254-942fcf7bee79-kube-api-access-2njjp\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703882 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703907 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.703962 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/07166244-e179-4d78-a254-942fcf7bee79-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.704004 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.705200 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.705247 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.705518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/07166244-e179-4d78-a254-942fcf7bee79-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.706813 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/07166244-e179-4d78-a254-942fcf7bee79-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.709846 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.709891 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c25b13bd4bcff2c06e8e9dc0ade8725c2ee3bc6d6a80507ef0c6769d5a6d3274/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.712119 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.721973 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/07166244-e179-4d78-a254-942fcf7bee79-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.731643 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2njjp\" (UniqueName: \"kubernetes.io/projected/07166244-e179-4d78-a254-942fcf7bee79-kube-api-access-2njjp\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.738349 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c","Type":"ContainerStarted","Data":"757e0ed0d91754c722d80f43d578638bda579b6360d4eda912709dc256b43f7f"} Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.753032 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-19d980f1-fa70-4b3c-9d5a-c34327d6b50e\") pod \"openstack-cell1-galera-0\" (UID: \"07166244-e179-4d78-a254-942fcf7bee79\") " pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.784209 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.792610 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.793690 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.796545 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5qs4v" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.796679 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.800538 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.802077 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.912551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kolla-config\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.912676 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.912926 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-config-data\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.912998 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:21 crc kubenswrapper[4867]: I1212 08:06:21.913118 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kube-api-access-88d65\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.014347 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-config-data\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.014925 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.014978 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kube-api-access-88d65\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.015054 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kolla-config\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.015081 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.015432 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-config-data\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.016551 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kolla-config\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.030600 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.034250 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.048969 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88d65\" (UniqueName: \"kubernetes.io/projected/01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a-kube-api-access-88d65\") pod \"memcached-0\" (UID: \"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a\") " pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.130759 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.289244 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 12 08:06:22 crc kubenswrapper[4867]: W1212 08:06:22.295305 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07166244_e179_4d78_a254_942fcf7bee79.slice/crio-b13ed4b05bf3859519a79d0943cc5def94bfe8ef7216508d237d7aee09693101 WatchSource:0}: Error finding container b13ed4b05bf3859519a79d0943cc5def94bfe8ef7216508d237d7aee09693101: Status 404 returned error can't find the container with id b13ed4b05bf3859519a79d0943cc5def94bfe8ef7216508d237d7aee09693101 Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.585074 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 12 08:06:22 crc kubenswrapper[4867]: W1212 08:06:22.592094 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01e9eb2a_e9ab_4fc6_980f_0cd5367e9c6a.slice/crio-b323ec7ff2eaf4ee2887ae5971ab029a8a714ec591ed5cf0625c319c6b5c4699 WatchSource:0}: Error finding container b323ec7ff2eaf4ee2887ae5971ab029a8a714ec591ed5cf0625c319c6b5c4699: Status 404 returned error can't find the container with id b323ec7ff2eaf4ee2887ae5971ab029a8a714ec591ed5cf0625c319c6b5c4699 Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.750986 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"07166244-e179-4d78-a254-942fcf7bee79","Type":"ContainerStarted","Data":"b13ed4b05bf3859519a79d0943cc5def94bfe8ef7216508d237d7aee09693101"} Dec 12 08:06:22 crc kubenswrapper[4867]: I1212 08:06:22.752702 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a","Type":"ContainerStarted","Data":"b323ec7ff2eaf4ee2887ae5971ab029a8a714ec591ed5cf0625c319c6b5c4699"} Dec 12 08:06:28 crc kubenswrapper[4867]: I1212 08:06:28.837815 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:06:28 crc kubenswrapper[4867]: E1212 08:06:28.838614 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:06:39 crc kubenswrapper[4867]: I1212 08:06:39.838182 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:06:39 crc kubenswrapper[4867]: E1212 08:06:39.838957 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.972549 4867 generic.go:334] "Generic (PLEG): container finished" podID="e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" containerID="5eab8466b76389d7d0c3d1138eb78431d6ee00c3cd00c53f16d263c5eba3a66d" exitCode=0 Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.973108 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" event={"ID":"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2","Type":"ContainerDied","Data":"5eab8466b76389d7d0c3d1138eb78431d6ee00c3cd00c53f16d263c5eba3a66d"} Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.975907 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c","Type":"ContainerStarted","Data":"a82015a7b8ff7ed43a390efe10daa1f2e85aad25cfa176f3f0434dcf657f91ac"} Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.987781 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"07166244-e179-4d78-a254-942fcf7bee79","Type":"ContainerStarted","Data":"980d3f6970689118974a73201339fd9b3b76b1c1839f377b81e2b92b5901124d"} Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.991091 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a","Type":"ContainerStarted","Data":"bfd0bdaba7e331ce180a611c1921c8cab98bd06f9487cfe02dc48eb95460f772"} Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.992750 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 12 08:06:46 crc kubenswrapper[4867]: I1212 08:06:46.997501 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerStarted","Data":"ed80a63b93616b7a7c662f722afb00461f86907cf49ccb3dc21498d4a7f760d1"} Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.009207 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerStarted","Data":"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0"} Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.011574 4867 generic.go:334] "Generic (PLEG): container finished" podID="36db5392-1167-463c-ab34-f97bcc329454" containerID="2f1b952a4d5d8bcb0c20f07e3bf278c4ff4fbbb1038ac35187b6bc078396d9a4" exitCode=0 Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.011788 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c47c755-pxw57" event={"ID":"36db5392-1167-463c-ab34-f97bcc329454","Type":"ContainerDied","Data":"2f1b952a4d5d8bcb0c20f07e3bf278c4ff4fbbb1038ac35187b6bc078396d9a4"} Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.013564 4867 generic.go:334] "Generic (PLEG): container finished" podID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerID="3f76c0929457d0d9fa12d770f38591e3b3aef00c424adacfb25cd9e4d4c11382" exitCode=0 Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.013620 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" event={"ID":"48f6e308-b8de-4401-817e-1b7201d6b13e","Type":"ContainerDied","Data":"3f76c0929457d0d9fa12d770f38591e3b3aef00c424adacfb25cd9e4d4c11382"} Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.015382 4867 generic.go:334] "Generic (PLEG): container finished" podID="eb845851-49b6-490d-b58b-e8ae0419d94c" containerID="e03b6fadb2af3b3ebf0d3ef48eb0fad1770f38450d609524135e482e35e942d1" exitCode=0 Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.015524 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" event={"ID":"eb845851-49b6-490d-b58b-e8ae0419d94c","Type":"ContainerDied","Data":"e03b6fadb2af3b3ebf0d3ef48eb0fad1770f38450d609524135e482e35e942d1"} Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.110849 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.258386663 podStartE2EDuration="26.11083014s" podCreationTimestamp="2025-12-12 08:06:21 +0000 UTC" firstStartedPulling="2025-12-12 08:06:22.596337165 +0000 UTC m=+4670.167718434" lastFinishedPulling="2025-12-12 08:06:45.448780642 +0000 UTC m=+4693.020161911" observedRunningTime="2025-12-12 08:06:47.108938354 +0000 UTC m=+4694.680319623" watchObservedRunningTime="2025-12-12 08:06:47.11083014 +0000 UTC m=+4694.682211409" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.271273 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.421722 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc\") pod \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.421796 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config\") pod \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.421888 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4npm\" (UniqueName: \"kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm\") pod \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\" (UID: \"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2\") " Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.425539 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm" (OuterVolumeSpecName: "kube-api-access-l4npm") pod "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" (UID: "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2"). InnerVolumeSpecName "kube-api-access-l4npm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.438821 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config" (OuterVolumeSpecName: "config") pod "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" (UID: "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.438833 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" (UID: "e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.447972 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.523040 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config\") pod \"eb845851-49b6-490d-b58b-e8ae0419d94c\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.523929 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbjpj\" (UniqueName: \"kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj\") pod \"eb845851-49b6-490d-b58b-e8ae0419d94c\" (UID: \"eb845851-49b6-490d-b58b-e8ae0419d94c\") " Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.524461 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.524529 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.524640 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4npm\" (UniqueName: \"kubernetes.io/projected/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2-kube-api-access-l4npm\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.526497 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj" (OuterVolumeSpecName: "kube-api-access-gbjpj") pod "eb845851-49b6-490d-b58b-e8ae0419d94c" (UID: "eb845851-49b6-490d-b58b-e8ae0419d94c"). InnerVolumeSpecName "kube-api-access-gbjpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.538846 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config" (OuterVolumeSpecName: "config") pod "eb845851-49b6-490d-b58b-e8ae0419d94c" (UID: "eb845851-49b6-490d-b58b-e8ae0419d94c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.626128 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb845851-49b6-490d-b58b-e8ae0419d94c-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:47 crc kubenswrapper[4867]: I1212 08:06:47.626160 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbjpj\" (UniqueName: \"kubernetes.io/projected/eb845851-49b6-490d-b58b-e8ae0419d94c-kube-api-access-gbjpj\") on node \"crc\" DevicePath \"\"" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.022466 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" event={"ID":"eb845851-49b6-490d-b58b-e8ae0419d94c","Type":"ContainerDied","Data":"a724406f4c3d79392679fa9133a92e5dcb7ff276148d1fb5da7458285b3cc22d"} Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.022512 4867 scope.go:117] "RemoveContainer" containerID="e03b6fadb2af3b3ebf0d3ef48eb0fad1770f38450d609524135e482e35e942d1" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.022621 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95485ddf-lq2xg" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.027588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" event={"ID":"e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2","Type":"ContainerDied","Data":"8204b65dd9ddbad2929d5bb89d19fb421c97a6425bc0e346686e70128b327cfd"} Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.027661 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6d89767f-2zpft" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.029651 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c47c755-pxw57" event={"ID":"36db5392-1167-463c-ab34-f97bcc329454","Type":"ContainerStarted","Data":"4f522289823c704fef13152c781a9eb6ea15b2acb689429105ef0196f5b330a4"} Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.029729 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.031550 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" event={"ID":"48f6e308-b8de-4401-817e-1b7201d6b13e","Type":"ContainerStarted","Data":"d9032b1aa643d003beb97a7c59d88091b595c317b26656fe4c1f24a9a0ea1352"} Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.063108 4867 scope.go:117] "RemoveContainer" containerID="5eab8466b76389d7d0c3d1138eb78431d6ee00c3cd00c53f16d263c5eba3a66d" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.066246 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" podStartSLOduration=4.135063901 podStartE2EDuration="31.066207969s" podCreationTimestamp="2025-12-12 08:06:17 +0000 UTC" firstStartedPulling="2025-12-12 08:06:18.636681669 +0000 UTC m=+4666.208062938" lastFinishedPulling="2025-12-12 08:06:45.567825737 +0000 UTC m=+4693.139207006" observedRunningTime="2025-12-12 08:06:48.057973716 +0000 UTC m=+4695.629354995" watchObservedRunningTime="2025-12-12 08:06:48.066207969 +0000 UTC m=+4695.637589238" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.079337 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55c47c755-pxw57" podStartSLOduration=4.999578426 podStartE2EDuration="31.07930785s" podCreationTimestamp="2025-12-12 08:06:17 +0000 UTC" firstStartedPulling="2025-12-12 08:06:19.475829642 +0000 UTC m=+4667.047210901" lastFinishedPulling="2025-12-12 08:06:45.555559056 +0000 UTC m=+4693.126940325" observedRunningTime="2025-12-12 08:06:48.074903482 +0000 UTC m=+4695.646284781" watchObservedRunningTime="2025-12-12 08:06:48.07930785 +0000 UTC m=+4695.650689129" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.120423 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.126857 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c95485ddf-lq2xg"] Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.151281 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.157103 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f6d89767f-2zpft"] Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.849774 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" path="/var/lib/kubelet/pods/e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2/volumes" Dec 12 08:06:48 crc kubenswrapper[4867]: I1212 08:06:48.851362 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb845851-49b6-490d-b58b-e8ae0419d94c" path="/var/lib/kubelet/pods/eb845851-49b6-490d-b58b-e8ae0419d94c/volumes" Dec 12 08:06:49 crc kubenswrapper[4867]: I1212 08:06:49.059244 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:50 crc kubenswrapper[4867]: I1212 08:06:50.068787 4867 generic.go:334] "Generic (PLEG): container finished" podID="07166244-e179-4d78-a254-942fcf7bee79" containerID="980d3f6970689118974a73201339fd9b3b76b1c1839f377b81e2b92b5901124d" exitCode=0 Dec 12 08:06:50 crc kubenswrapper[4867]: I1212 08:06:50.068865 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"07166244-e179-4d78-a254-942fcf7bee79","Type":"ContainerDied","Data":"980d3f6970689118974a73201339fd9b3b76b1c1839f377b81e2b92b5901124d"} Dec 12 08:06:50 crc kubenswrapper[4867]: I1212 08:06:50.070855 4867 generic.go:334] "Generic (PLEG): container finished" podID="0583cfff-0ec3-4fad-84de-3c9f6f3dd49c" containerID="a82015a7b8ff7ed43a390efe10daa1f2e85aad25cfa176f3f0434dcf657f91ac" exitCode=0 Dec 12 08:06:50 crc kubenswrapper[4867]: I1212 08:06:50.070876 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c","Type":"ContainerDied","Data":"a82015a7b8ff7ed43a390efe10daa1f2e85aad25cfa176f3f0434dcf657f91ac"} Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.081317 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0583cfff-0ec3-4fad-84de-3c9f6f3dd49c","Type":"ContainerStarted","Data":"eaa36a45a58e27145a38699c23c95abf65bde28275e1d96441bdb12fe84e49bf"} Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.083801 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"07166244-e179-4d78-a254-942fcf7bee79","Type":"ContainerStarted","Data":"8c2e372824cfe968c75904942bb7822c833f5cdffbf28f01879861e0a04bb8d1"} Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.107362 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.469292442 podStartE2EDuration="33.107339212s" podCreationTimestamp="2025-12-12 08:06:18 +0000 UTC" firstStartedPulling="2025-12-12 08:06:20.856321293 +0000 UTC m=+4668.427702562" lastFinishedPulling="2025-12-12 08:06:45.494368063 +0000 UTC m=+4693.065749332" observedRunningTime="2025-12-12 08:06:51.100950725 +0000 UTC m=+4698.672332004" watchObservedRunningTime="2025-12-12 08:06:51.107339212 +0000 UTC m=+4698.678720481" Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.118304 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.8524044570000004 podStartE2EDuration="31.118283721s" podCreationTimestamp="2025-12-12 08:06:20 +0000 UTC" firstStartedPulling="2025-12-12 08:06:22.307327995 +0000 UTC m=+4669.878709264" lastFinishedPulling="2025-12-12 08:06:45.573207259 +0000 UTC m=+4693.144588528" observedRunningTime="2025-12-12 08:06:51.117279496 +0000 UTC m=+4698.688660775" watchObservedRunningTime="2025-12-12 08:06:51.118283721 +0000 UTC m=+4698.689664990" Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.785301 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:51 crc kubenswrapper[4867]: I1212 08:06:51.785344 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 12 08:06:52 crc kubenswrapper[4867]: I1212 08:06:52.132235 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 12 08:06:52 crc kubenswrapper[4867]: I1212 08:06:52.682506 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:06:53 crc kubenswrapper[4867]: I1212 08:06:53.227135 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:06:53 crc kubenswrapper[4867]: I1212 08:06:53.282357 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:06:53 crc kubenswrapper[4867]: I1212 08:06:53.282560 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="dnsmasq-dns" containerID="cri-o://d9032b1aa643d003beb97a7c59d88091b595c317b26656fe4c1f24a9a0ea1352" gracePeriod=10 Dec 12 08:06:54 crc kubenswrapper[4867]: I1212 08:06:54.838274 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:06:54 crc kubenswrapper[4867]: E1212 08:06:54.838781 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:06:55 crc kubenswrapper[4867]: I1212 08:06:55.124770 4867 generic.go:334] "Generic (PLEG): container finished" podID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerID="d9032b1aa643d003beb97a7c59d88091b595c317b26656fe4c1f24a9a0ea1352" exitCode=0 Dec 12 08:06:55 crc kubenswrapper[4867]: I1212 08:06:55.124828 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" event={"ID":"48f6e308-b8de-4401-817e-1b7201d6b13e","Type":"ContainerDied","Data":"d9032b1aa643d003beb97a7c59d88091b595c317b26656fe4c1f24a9a0ea1352"} Dec 12 08:06:57 crc kubenswrapper[4867]: I1212 08:06:57.681446 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.240:5353: connect: connection refused" Dec 12 08:07:00 crc kubenswrapper[4867]: I1212 08:07:00.265299 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 12 08:07:00 crc kubenswrapper[4867]: I1212 08:07:00.265973 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.146717 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.176236 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" event={"ID":"48f6e308-b8de-4401-817e-1b7201d6b13e","Type":"ContainerDied","Data":"8087338e061c95283ce3c115f7734bd000372b55e2c61e9257ed881e0092a3f6"} Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.176281 4867 scope.go:117] "RemoveContainer" containerID="d9032b1aa643d003beb97a7c59d88091b595c317b26656fe4c1f24a9a0ea1352" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.176405 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f5bc65899-mb2tr" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.198041 4867 scope.go:117] "RemoveContainer" containerID="3f76c0929457d0d9fa12d770f38591e3b3aef00c424adacfb25cd9e4d4c11382" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.257810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc\") pod \"48f6e308-b8de-4401-817e-1b7201d6b13e\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.257977 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpp9s\" (UniqueName: \"kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s\") pod \"48f6e308-b8de-4401-817e-1b7201d6b13e\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.258016 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config\") pod \"48f6e308-b8de-4401-817e-1b7201d6b13e\" (UID: \"48f6e308-b8de-4401-817e-1b7201d6b13e\") " Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.263606 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s" (OuterVolumeSpecName: "kube-api-access-jpp9s") pod "48f6e308-b8de-4401-817e-1b7201d6b13e" (UID: "48f6e308-b8de-4401-817e-1b7201d6b13e"). InnerVolumeSpecName "kube-api-access-jpp9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.295982 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48f6e308-b8de-4401-817e-1b7201d6b13e" (UID: "48f6e308-b8de-4401-817e-1b7201d6b13e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.297661 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config" (OuterVolumeSpecName: "config") pod "48f6e308-b8de-4401-817e-1b7201d6b13e" (UID: "48f6e308-b8de-4401-817e-1b7201d6b13e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.359666 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpp9s\" (UniqueName: \"kubernetes.io/projected/48f6e308-b8de-4401-817e-1b7201d6b13e-kube-api-access-jpp9s\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.359706 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.359716 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48f6e308-b8de-4401-817e-1b7201d6b13e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.518931 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:07:01 crc kubenswrapper[4867]: I1212 08:07:01.524076 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f5bc65899-mb2tr"] Dec 12 08:07:01 crc kubenswrapper[4867]: E1212 08:07:01.750212 4867 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.66:52962->38.102.83.66:33599: write tcp 38.102.83.66:52962->38.102.83.66:33599: write: broken pipe Dec 12 08:07:02 crc kubenswrapper[4867]: I1212 08:07:02.849599 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" path="/var/lib/kubelet/pods/48f6e308-b8de-4401-817e-1b7201d6b13e/volumes" Dec 12 08:07:03 crc kubenswrapper[4867]: I1212 08:07:03.895655 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 12 08:07:03 crc kubenswrapper[4867]: I1212 08:07:03.964127 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 12 08:07:04 crc kubenswrapper[4867]: I1212 08:07:04.939444 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 12 08:07:05 crc kubenswrapper[4867]: I1212 08:07:05.015076 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 12 08:07:05 crc kubenswrapper[4867]: I1212 08:07:05.838290 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:07:05 crc kubenswrapper[4867]: E1212 08:07:05.838511 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:07:17 crc kubenswrapper[4867]: I1212 08:07:17.838902 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:07:17 crc kubenswrapper[4867]: E1212 08:07:17.839894 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:07:18 crc kubenswrapper[4867]: E1212 08:07:18.680668 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod734c7cce_d153_4239_a3ff_e876fac8452f.slice/crio-9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:07:19 crc kubenswrapper[4867]: I1212 08:07:19.303333 4867 generic.go:334] "Generic (PLEG): container finished" podID="2300b08d-c21e-4cf2-bca3-97440b179709" containerID="ed80a63b93616b7a7c662f722afb00461f86907cf49ccb3dc21498d4a7f760d1" exitCode=0 Dec 12 08:07:19 crc kubenswrapper[4867]: I1212 08:07:19.303420 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerDied","Data":"ed80a63b93616b7a7c662f722afb00461f86907cf49ccb3dc21498d4a7f760d1"} Dec 12 08:07:19 crc kubenswrapper[4867]: I1212 08:07:19.304710 4867 generic.go:334] "Generic (PLEG): container finished" podID="734c7cce-d153-4239-a3ff-e876fac8452f" containerID="9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0" exitCode=0 Dec 12 08:07:19 crc kubenswrapper[4867]: I1212 08:07:19.304767 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerDied","Data":"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0"} Dec 12 08:07:20 crc kubenswrapper[4867]: I1212 08:07:20.316772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerStarted","Data":"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f"} Dec 12 08:07:20 crc kubenswrapper[4867]: I1212 08:07:20.318146 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:20 crc kubenswrapper[4867]: I1212 08:07:20.345883 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.070145713 podStartE2EDuration="1m3.345866357s" podCreationTimestamp="2025-12-12 08:06:17 +0000 UTC" firstStartedPulling="2025-12-12 08:06:20.219405727 +0000 UTC m=+4667.790786996" lastFinishedPulling="2025-12-12 08:06:45.495126371 +0000 UTC m=+4693.066507640" observedRunningTime="2025-12-12 08:07:20.33910038 +0000 UTC m=+4727.910481659" watchObservedRunningTime="2025-12-12 08:07:20.345866357 +0000 UTC m=+4727.917247616" Dec 12 08:07:21 crc kubenswrapper[4867]: I1212 08:07:21.325067 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerStarted","Data":"d96f2f9c86ced7f50881c60c5e46372c2c17c7b40a35a82b7c6123005a6cd567"} Dec 12 08:07:21 crc kubenswrapper[4867]: I1212 08:07:21.325604 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 12 08:07:21 crc kubenswrapper[4867]: I1212 08:07:21.349984 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.732299708 podStartE2EDuration="1m4.349959692s" podCreationTimestamp="2025-12-12 08:06:17 +0000 UTC" firstStartedPulling="2025-12-12 08:06:19.617584514 +0000 UTC m=+4667.188965783" lastFinishedPulling="2025-12-12 08:06:43.235244498 +0000 UTC m=+4690.806625767" observedRunningTime="2025-12-12 08:07:21.348795114 +0000 UTC m=+4728.920176393" watchObservedRunningTime="2025-12-12 08:07:21.349959692 +0000 UTC m=+4728.921340961" Dec 12 08:07:29 crc kubenswrapper[4867]: I1212 08:07:29.341187 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.243:5671: connect: connection refused" Dec 12 08:07:29 crc kubenswrapper[4867]: I1212 08:07:29.838752 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:07:29 crc kubenswrapper[4867]: E1212 08:07:29.839007 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:07:38 crc kubenswrapper[4867]: I1212 08:07:38.785434 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 12 08:07:39 crc kubenswrapper[4867]: I1212 08:07:39.340403 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:44 crc kubenswrapper[4867]: I1212 08:07:44.837790 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:07:44 crc kubenswrapper[4867]: E1212 08:07:44.838555 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.066149 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:07:45 crc kubenswrapper[4867]: E1212 08:07:45.066860 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="dnsmasq-dns" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.066882 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="dnsmasq-dns" Dec 12 08:07:45 crc kubenswrapper[4867]: E1212 08:07:45.066909 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb845851-49b6-490d-b58b-e8ae0419d94c" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.066918 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb845851-49b6-490d-b58b-e8ae0419d94c" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: E1212 08:07:45.066941 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.066949 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: E1212 08:07:45.066962 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.066970 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.067139 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb845851-49b6-490d-b58b-e8ae0419d94c" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.067174 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f6e308-b8de-4401-817e-1b7201d6b13e" containerName="dnsmasq-dns" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.067186 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f22a8a-bbc5-42da-a8e7-c5a0edd79fc2" containerName="init" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.068164 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.095615 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.162662 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.163083 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdbn\" (UniqueName: \"kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.163280 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.264941 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.265314 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.265499 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdbn\" (UniqueName: \"kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.266117 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.266412 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.284206 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdbn\" (UniqueName: \"kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn\") pod \"dnsmasq-dns-76958896f7-gv5hr\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.385927 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.782070 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:45 crc kubenswrapper[4867]: I1212 08:07:45.955343 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:07:45 crc kubenswrapper[4867]: W1212 08:07:45.959164 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45b18049_f199_4ad0_ad94_64e06149d167.slice/crio-ee6e504d4bc9a0e755579848ab8855698ab16a72f701d17aadeffcf337d131aa WatchSource:0}: Error finding container ee6e504d4bc9a0e755579848ab8855698ab16a72f701d17aadeffcf337d131aa: Status 404 returned error can't find the container with id ee6e504d4bc9a0e755579848ab8855698ab16a72f701d17aadeffcf337d131aa Dec 12 08:07:46 crc kubenswrapper[4867]: I1212 08:07:46.439639 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:46 crc kubenswrapper[4867]: I1212 08:07:46.521125 4867 generic.go:334] "Generic (PLEG): container finished" podID="45b18049-f199-4ad0-ad94-64e06149d167" containerID="34c1f2e171b02a46ec310aed4c241d113d2ea43dc63d7da2c5451504ca72972c" exitCode=0 Dec 12 08:07:46 crc kubenswrapper[4867]: I1212 08:07:46.521176 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" event={"ID":"45b18049-f199-4ad0-ad94-64e06149d167","Type":"ContainerDied","Data":"34c1f2e171b02a46ec310aed4c241d113d2ea43dc63d7da2c5451504ca72972c"} Dec 12 08:07:46 crc kubenswrapper[4867]: I1212 08:07:46.521204 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" event={"ID":"45b18049-f199-4ad0-ad94-64e06149d167","Type":"ContainerStarted","Data":"ee6e504d4bc9a0e755579848ab8855698ab16a72f701d17aadeffcf337d131aa"} Dec 12 08:07:47 crc kubenswrapper[4867]: I1212 08:07:47.529735 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" event={"ID":"45b18049-f199-4ad0-ad94-64e06149d167","Type":"ContainerStarted","Data":"e9313c3280f84f2be678dcb7b1936ccd5cee1ecc08176b58bd858e60b6fbbef7"} Dec 12 08:07:47 crc kubenswrapper[4867]: I1212 08:07:47.530111 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:47 crc kubenswrapper[4867]: I1212 08:07:47.549641 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" podStartSLOduration=2.549586859 podStartE2EDuration="2.549586859s" podCreationTimestamp="2025-12-12 08:07:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:07:47.547819825 +0000 UTC m=+4755.119201094" watchObservedRunningTime="2025-12-12 08:07:47.549586859 +0000 UTC m=+4755.120968128" Dec 12 08:07:50 crc kubenswrapper[4867]: I1212 08:07:50.172908 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="rabbitmq" containerID="cri-o://d96f2f9c86ced7f50881c60c5e46372c2c17c7b40a35a82b7c6123005a6cd567" gracePeriod=604796 Dec 12 08:07:50 crc kubenswrapper[4867]: I1212 08:07:50.612849 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="rabbitmq" containerID="cri-o://2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f" gracePeriod=604796 Dec 12 08:07:55 crc kubenswrapper[4867]: I1212 08:07:55.387464 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:07:55 crc kubenswrapper[4867]: I1212 08:07:55.444063 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:07:55 crc kubenswrapper[4867]: I1212 08:07:55.446450 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55c47c755-pxw57" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="dnsmasq-dns" containerID="cri-o://4f522289823c704fef13152c781a9eb6ea15b2acb689429105ef0196f5b330a4" gracePeriod=10 Dec 12 08:07:55 crc kubenswrapper[4867]: I1212 08:07:55.609604 4867 generic.go:334] "Generic (PLEG): container finished" podID="36db5392-1167-463c-ab34-f97bcc329454" containerID="4f522289823c704fef13152c781a9eb6ea15b2acb689429105ef0196f5b330a4" exitCode=0 Dec 12 08:07:55 crc kubenswrapper[4867]: I1212 08:07:55.610082 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c47c755-pxw57" event={"ID":"36db5392-1167-463c-ab34-f97bcc329454","Type":"ContainerDied","Data":"4f522289823c704fef13152c781a9eb6ea15b2acb689429105ef0196f5b330a4"} Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.030881 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.142998 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc\") pod \"36db5392-1167-463c-ab34-f97bcc329454\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.143060 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config\") pod \"36db5392-1167-463c-ab34-f97bcc329454\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.143103 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkbkr\" (UniqueName: \"kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr\") pod \"36db5392-1167-463c-ab34-f97bcc329454\" (UID: \"36db5392-1167-463c-ab34-f97bcc329454\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.155489 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr" (OuterVolumeSpecName: "kube-api-access-gkbkr") pod "36db5392-1167-463c-ab34-f97bcc329454" (UID: "36db5392-1167-463c-ab34-f97bcc329454"). InnerVolumeSpecName "kube-api-access-gkbkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.186615 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "36db5392-1167-463c-ab34-f97bcc329454" (UID: "36db5392-1167-463c-ab34-f97bcc329454"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.186764 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config" (OuterVolumeSpecName: "config") pod "36db5392-1167-463c-ab34-f97bcc329454" (UID: "36db5392-1167-463c-ab34-f97bcc329454"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.244425 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.244461 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36db5392-1167-463c-ab34-f97bcc329454-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.244471 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkbkr\" (UniqueName: \"kubernetes.io/projected/36db5392-1167-463c-ab34-f97bcc329454-kube-api-access-gkbkr\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.628945 4867 generic.go:334] "Generic (PLEG): container finished" podID="2300b08d-c21e-4cf2-bca3-97440b179709" containerID="d96f2f9c86ced7f50881c60c5e46372c2c17c7b40a35a82b7c6123005a6cd567" exitCode=0 Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.629008 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerDied","Data":"d96f2f9c86ced7f50881c60c5e46372c2c17c7b40a35a82b7c6123005a6cd567"} Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.630630 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55c47c755-pxw57" event={"ID":"36db5392-1167-463c-ab34-f97bcc329454","Type":"ContainerDied","Data":"0b7ebabfeda3511e0a73117b2c248842bb0d044d957aae5a9597bf35e92a1dcb"} Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.630663 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c47c755-pxw57" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.630682 4867 scope.go:117] "RemoveContainer" containerID="4f522289823c704fef13152c781a9eb6ea15b2acb689429105ef0196f5b330a4" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.651526 4867 scope.go:117] "RemoveContainer" containerID="2f1b952a4d5d8bcb0c20f07e3bf278c4ff4fbbb1038ac35187b6bc078396d9a4" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.674158 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.679619 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55c47c755-pxw57"] Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.806404 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.846510 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36db5392-1167-463c-ab34-f97bcc329454" path="/var/lib/kubelet/pods/36db5392-1167-463c-ab34-f97bcc329454/volumes" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953551 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hklsv\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953636 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953678 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953731 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953843 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953879 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953904 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953946 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.953969 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.954000 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie\") pod \"2300b08d-c21e-4cf2-bca3-97440b179709\" (UID: \"2300b08d-c21e-4cf2-bca3-97440b179709\") " Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.955147 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.955627 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.955785 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.960123 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.965840 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.974172 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv" (OuterVolumeSpecName: "kube-api-access-hklsv") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "kube-api-access-hklsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.976327 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info" (OuterVolumeSpecName: "pod-info") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.979441 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee" (OuterVolumeSpecName: "persistence") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 08:07:56 crc kubenswrapper[4867]: I1212 08:07:56.981739 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data" (OuterVolumeSpecName: "config-data") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.020589 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf" (OuterVolumeSpecName: "server-conf") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.049104 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2300b08d-c21e-4cf2-bca3-97440b179709" (UID: "2300b08d-c21e-4cf2-bca3-97440b179709"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057281 4867 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-server-conf\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057316 4867 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057328 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057339 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2300b08d-c21e-4cf2-bca3-97440b179709-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057349 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057359 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hklsv\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-kube-api-access-hklsv\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057366 4867 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2300b08d-c21e-4cf2-bca3-97440b179709-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057374 4867 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2300b08d-c21e-4cf2-bca3-97440b179709-pod-info\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057382 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057391 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2300b08d-c21e-4cf2-bca3-97440b179709-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.057430 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") on node \"crc\" " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.074538 4867 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.074670 4867 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee") on node "crc" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.125667 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.161268 4867 reconciler_common.go:293] "Volume detached for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262584 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262643 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262686 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262721 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262752 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.262955 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263013 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263054 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bsjp\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263093 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263114 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263148 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data\") pod \"734c7cce-d153-4239-a3ff-e876fac8452f\" (UID: \"734c7cce-d153-4239-a3ff-e876fac8452f\") " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263272 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.263424 4867 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.264067 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.264545 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.267102 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.267978 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info" (OuterVolumeSpecName: "pod-info") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.268144 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.272530 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp" (OuterVolumeSpecName: "kube-api-access-2bsjp") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "kube-api-access-2bsjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.287856 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1" (OuterVolumeSpecName: "persistence") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.296921 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data" (OuterVolumeSpecName: "config-data") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.302950 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf" (OuterVolumeSpecName: "server-conf") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.353640 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "734c7cce-d153-4239-a3ff-e876fac8452f" (UID: "734c7cce-d153-4239-a3ff-e876fac8452f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364540 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364592 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bsjp\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-kube-api-access-2bsjp\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364610 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364622 4867 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/734c7cce-d153-4239-a3ff-e876fac8452f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364635 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364649 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364659 4867 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/734c7cce-d153-4239-a3ff-e876fac8452f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364670 4867 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/734c7cce-d153-4239-a3ff-e876fac8452f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364680 4867 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/734c7cce-d153-4239-a3ff-e876fac8452f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.364715 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") on node \"crc\" " Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.386284 4867 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.386446 4867 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1") on node "crc" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.465972 4867 reconciler_common.go:293] "Volume detached for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") on node \"crc\" DevicePath \"\"" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.640138 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2300b08d-c21e-4cf2-bca3-97440b179709","Type":"ContainerDied","Data":"9e2c3a24f1d4ac6ee9852749a445a8d021dd4c9efd389e9d82e7310ec3a75f02"} Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.640186 4867 scope.go:117] "RemoveContainer" containerID="d96f2f9c86ced7f50881c60c5e46372c2c17c7b40a35a82b7c6123005a6cd567" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.640286 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.649729 4867 generic.go:334] "Generic (PLEG): container finished" podID="734c7cce-d153-4239-a3ff-e876fac8452f" containerID="2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f" exitCode=0 Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.649774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerDied","Data":"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f"} Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.649803 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"734c7cce-d153-4239-a3ff-e876fac8452f","Type":"ContainerDied","Data":"f7c31e3aa2319e09d3b415f6f2b367bedfa320d3674b4f923e6c1031f962386c"} Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.649821 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.677319 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.681039 4867 scope.go:117] "RemoveContainer" containerID="ed80a63b93616b7a7c662f722afb00461f86907cf49ccb3dc21498d4a7f760d1" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.681750 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.692711 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.705656 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714218 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714591 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714612 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714636 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714647 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714659 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="setup-container" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714668 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="setup-container" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714683 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="dnsmasq-dns" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714690 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="dnsmasq-dns" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714702 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="setup-container" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714710 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="setup-container" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.714733 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="init" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714741 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="init" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714924 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714948 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="36db5392-1167-463c-ab34-f97bcc329454" containerName="dnsmasq-dns" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.714965 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" containerName="rabbitmq" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.715977 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.719407 4867 scope.go:117] "RemoveContainer" containerID="2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.725694 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.727722 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.727771 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.727870 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.727972 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wklzn" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.728073 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.728126 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.728169 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.747060 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.748571 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.759456 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.759710 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.759907 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.760058 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-cjs79" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.760194 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.760390 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.760530 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.771496 4867 scope.go:117] "RemoveContainer" containerID="9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.783355 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.820137 4867 scope.go:117] "RemoveContainer" containerID="2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.821802 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f\": container with ID starting with 2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f not found: ID does not exist" containerID="2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.821863 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f"} err="failed to get container status \"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f\": rpc error: code = NotFound desc = could not find container \"2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f\": container with ID starting with 2dc64f80c514c63347b0a4154dbb501a9e523c74b04c236816e140d933432c0f not found: ID does not exist" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.821887 4867 scope.go:117] "RemoveContainer" containerID="9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.828635 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0\": container with ID starting with 9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0 not found: ID does not exist" containerID="9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.828691 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0"} err="failed to get container status \"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0\": rpc error: code = NotFound desc = could not find container \"9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0\": container with ID starting with 9d9420f3b8d2fe39d6fff37a03edc22c64005965b3f4c9df1f83611b9a1467b0 not found: ID does not exist" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.838204 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:07:57 crc kubenswrapper[4867]: E1212 08:07:57.838454 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874064 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml29z\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-kube-api-access-ml29z\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874113 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874131 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eab4604e-3919-4f11-89ea-057524b8bfb9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874155 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874179 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874202 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-config-data\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874255 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874284 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874307 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874323 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae1a3a41-6283-4343-9d69-3fce5acbf570-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874340 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvhh\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-kube-api-access-xmvhh\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874365 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eab4604e-3919-4f11-89ea-057524b8bfb9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874380 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874397 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874422 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874439 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874462 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae1a3a41-6283-4343-9d69-3fce5acbf570-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874481 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874504 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874522 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874546 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.874564 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976120 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976187 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976206 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml29z\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-kube-api-access-ml29z\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976280 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976315 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eab4604e-3919-4f11-89ea-057524b8bfb9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976367 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976397 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976877 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.976949 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.977003 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-config-data\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.977100 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.977133 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.977157 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.977809 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-config-data\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978336 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978429 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eab4604e-3919-4f11-89ea-057524b8bfb9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978477 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae1a3a41-6283-4343-9d69-3fce5acbf570-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978504 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvhh\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-kube-api-access-xmvhh\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978544 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eab4604e-3919-4f11-89ea-057524b8bfb9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978567 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978600 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978649 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978691 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978739 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae1a3a41-6283-4343-9d69-3fce5acbf570-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978770 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978791 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978826 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ade0757b040b265f114a1895a4de61d5d63bae711e8c14acb63c5e5d4e3b9f57/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978842 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.978872 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.979601 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.979727 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.980213 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.980575 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.981146 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ae1a3a41-6283-4343-9d69-3fce5acbf570-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.981181 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eab4604e-3919-4f11-89ea-057524b8bfb9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.981868 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.983609 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.983642 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f952049225bdbbdac120d8064eb4b6b35c89d69f6e2090c2cba213d8afde8e90/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.983818 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.984066 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.993631 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ae1a3a41-6283-4343-9d69-3fce5acbf570-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.993780 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ae1a3a41-6283-4343-9d69-3fce5acbf570-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.993901 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eab4604e-3919-4f11-89ea-057524b8bfb9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.996374 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml29z\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-kube-api-access-ml29z\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.998664 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvhh\" (UniqueName: \"kubernetes.io/projected/eab4604e-3919-4f11-89ea-057524b8bfb9-kube-api-access-xmvhh\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:57 crc kubenswrapper[4867]: I1212 08:07:57.999590 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ae1a3a41-6283-4343-9d69-3fce5acbf570-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.020779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bb0eb180-a053-4800-8e43-f8cde8f9c4ee\") pod \"rabbitmq-server-0\" (UID: \"eab4604e-3919-4f11-89ea-057524b8bfb9\") " pod="openstack/rabbitmq-server-0" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.046692 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162a3b17-33e6-472e-b0f3-9a264f93d0c1\") pod \"rabbitmq-cell1-server-0\" (UID: \"ae1a3a41-6283-4343-9d69-3fce5acbf570\") " pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.073809 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.107910 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.375763 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.527995 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 12 08:07:58 crc kubenswrapper[4867]: W1212 08:07:58.535494 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab4604e_3919_4f11_89ea_057524b8bfb9.slice/crio-7aaad0cd3d25620eba045d5c5b921c588fd250beda65aee5034cf392e28eefa3 WatchSource:0}: Error finding container 7aaad0cd3d25620eba045d5c5b921c588fd250beda65aee5034cf392e28eefa3: Status 404 returned error can't find the container with id 7aaad0cd3d25620eba045d5c5b921c588fd250beda65aee5034cf392e28eefa3 Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.659123 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ae1a3a41-6283-4343-9d69-3fce5acbf570","Type":"ContainerStarted","Data":"2626be5bd4d2c6ae6545e99c20ed7dd91ecc9af7694aaee36912f1cc407f6e74"} Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.663606 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eab4604e-3919-4f11-89ea-057524b8bfb9","Type":"ContainerStarted","Data":"7aaad0cd3d25620eba045d5c5b921c588fd250beda65aee5034cf392e28eefa3"} Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.847365 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2300b08d-c21e-4cf2-bca3-97440b179709" path="/var/lib/kubelet/pods/2300b08d-c21e-4cf2-bca3-97440b179709/volumes" Dec 12 08:07:58 crc kubenswrapper[4867]: I1212 08:07:58.848119 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="734c7cce-d153-4239-a3ff-e876fac8452f" path="/var/lib/kubelet/pods/734c7cce-d153-4239-a3ff-e876fac8452f/volumes" Dec 12 08:07:59 crc kubenswrapper[4867]: I1212 08:07:59.673197 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ae1a3a41-6283-4343-9d69-3fce5acbf570","Type":"ContainerStarted","Data":"e984e02de1aeaa794292c46766c4424adbb14b6cb43b418f35a0e52b59912ff9"} Dec 12 08:07:59 crc kubenswrapper[4867]: I1212 08:07:59.675455 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eab4604e-3919-4f11-89ea-057524b8bfb9","Type":"ContainerStarted","Data":"a4b0477e2e8d92816c86aca9cbc173e4d42b7c5eef0b59dd137e6ab82a5b9063"} Dec 12 08:08:10 crc kubenswrapper[4867]: I1212 08:08:10.838431 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:08:10 crc kubenswrapper[4867]: E1212 08:08:10.839260 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:08:23 crc kubenswrapper[4867]: I1212 08:08:23.838266 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:08:23 crc kubenswrapper[4867]: E1212 08:08:23.838912 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:08:31 crc kubenswrapper[4867]: I1212 08:08:31.911637 4867 generic.go:334] "Generic (PLEG): container finished" podID="ae1a3a41-6283-4343-9d69-3fce5acbf570" containerID="e984e02de1aeaa794292c46766c4424adbb14b6cb43b418f35a0e52b59912ff9" exitCode=0 Dec 12 08:08:31 crc kubenswrapper[4867]: I1212 08:08:31.911736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ae1a3a41-6283-4343-9d69-3fce5acbf570","Type":"ContainerDied","Data":"e984e02de1aeaa794292c46766c4424adbb14b6cb43b418f35a0e52b59912ff9"} Dec 12 08:08:31 crc kubenswrapper[4867]: I1212 08:08:31.914048 4867 generic.go:334] "Generic (PLEG): container finished" podID="eab4604e-3919-4f11-89ea-057524b8bfb9" containerID="a4b0477e2e8d92816c86aca9cbc173e4d42b7c5eef0b59dd137e6ab82a5b9063" exitCode=0 Dec 12 08:08:31 crc kubenswrapper[4867]: I1212 08:08:31.914094 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eab4604e-3919-4f11-89ea-057524b8bfb9","Type":"ContainerDied","Data":"a4b0477e2e8d92816c86aca9cbc173e4d42b7c5eef0b59dd137e6ab82a5b9063"} Dec 12 08:08:32 crc kubenswrapper[4867]: I1212 08:08:32.923648 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ae1a3a41-6283-4343-9d69-3fce5acbf570","Type":"ContainerStarted","Data":"72785ab1cb97e70f50d0c98a0d1d19b90ed79082499773fb023d816c20bc6aed"} Dec 12 08:08:32 crc kubenswrapper[4867]: I1212 08:08:32.925121 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:08:32 crc kubenswrapper[4867]: I1212 08:08:32.928050 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eab4604e-3919-4f11-89ea-057524b8bfb9","Type":"ContainerStarted","Data":"45c08c9e6ad8f2d121813a6ae8b2a8e6cd146a4b58146fe37d2da6bf3c3fe047"} Dec 12 08:08:32 crc kubenswrapper[4867]: I1212 08:08:32.928757 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 12 08:08:33 crc kubenswrapper[4867]: I1212 08:08:33.000392 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.000367295 podStartE2EDuration="36.000367295s" podCreationTimestamp="2025-12-12 08:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:08:32.962678829 +0000 UTC m=+4800.534060098" watchObservedRunningTime="2025-12-12 08:08:33.000367295 +0000 UTC m=+4800.571748574" Dec 12 08:08:33 crc kubenswrapper[4867]: I1212 08:08:33.004577 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.004567028 podStartE2EDuration="36.004567028s" podCreationTimestamp="2025-12-12 08:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:08:33.000992451 +0000 UTC m=+4800.572373720" watchObservedRunningTime="2025-12-12 08:08:33.004567028 +0000 UTC m=+4800.575948297" Dec 12 08:08:38 crc kubenswrapper[4867]: I1212 08:08:38.838249 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:08:38 crc kubenswrapper[4867]: E1212 08:08:38.839110 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:08:48 crc kubenswrapper[4867]: I1212 08:08:48.077421 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 12 08:08:48 crc kubenswrapper[4867]: I1212 08:08:48.111437 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.631253 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.632808 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.637685 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zg96v" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.639940 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.706786 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg9mk\" (UniqueName: \"kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk\") pod \"mariadb-client-1-default\" (UID: \"76adfb82-2076-4040-a838-341d9595a9a1\") " pod="openstack/mariadb-client-1-default" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.808193 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg9mk\" (UniqueName: \"kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk\") pod \"mariadb-client-1-default\" (UID: \"76adfb82-2076-4040-a838-341d9595a9a1\") " pod="openstack/mariadb-client-1-default" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.831497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg9mk\" (UniqueName: \"kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk\") pod \"mariadb-client-1-default\" (UID: \"76adfb82-2076-4040-a838-341d9595a9a1\") " pod="openstack/mariadb-client-1-default" Dec 12 08:08:52 crc kubenswrapper[4867]: I1212 08:08:52.961014 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 12 08:08:53 crc kubenswrapper[4867]: I1212 08:08:53.442962 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 12 08:08:53 crc kubenswrapper[4867]: I1212 08:08:53.452590 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:08:53 crc kubenswrapper[4867]: I1212 08:08:53.837854 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:08:53 crc kubenswrapper[4867]: E1212 08:08:53.838210 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:08:54 crc kubenswrapper[4867]: I1212 08:08:54.133737 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"76adfb82-2076-4040-a838-341d9595a9a1","Type":"ContainerStarted","Data":"2b70e0483a7ca0b2160ffc2e1ac5c4b7b6b19aa457253ecdb092c883fde117af"} Dec 12 08:08:55 crc kubenswrapper[4867]: I1212 08:08:55.141948 4867 generic.go:334] "Generic (PLEG): container finished" podID="76adfb82-2076-4040-a838-341d9595a9a1" containerID="10c7f6697285f3a68eea37a950dac79f44458cfef00879c19faa8bad0a331a98" exitCode=0 Dec 12 08:08:55 crc kubenswrapper[4867]: I1212 08:08:55.142222 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"76adfb82-2076-4040-a838-341d9595a9a1","Type":"ContainerDied","Data":"10c7f6697285f3a68eea37a950dac79f44458cfef00879c19faa8bad0a331a98"} Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.469763 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.494127 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_76adfb82-2076-4040-a838-341d9595a9a1/mariadb-client-1-default/0.log" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.520788 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.525284 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.669069 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg9mk\" (UniqueName: \"kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk\") pod \"76adfb82-2076-4040-a838-341d9595a9a1\" (UID: \"76adfb82-2076-4040-a838-341d9595a9a1\") " Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.684064 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk" (OuterVolumeSpecName: "kube-api-access-lg9mk") pod "76adfb82-2076-4040-a838-341d9595a9a1" (UID: "76adfb82-2076-4040-a838-341d9595a9a1"). InnerVolumeSpecName "kube-api-access-lg9mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.771549 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg9mk\" (UniqueName: \"kubernetes.io/projected/76adfb82-2076-4040-a838-341d9595a9a1-kube-api-access-lg9mk\") on node \"crc\" DevicePath \"\"" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.846364 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76adfb82-2076-4040-a838-341d9595a9a1" path="/var/lib/kubelet/pods/76adfb82-2076-4040-a838-341d9595a9a1/volumes" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.907672 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 12 08:08:56 crc kubenswrapper[4867]: E1212 08:08:56.908079 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76adfb82-2076-4040-a838-341d9595a9a1" containerName="mariadb-client-1-default" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.908101 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="76adfb82-2076-4040-a838-341d9595a9a1" containerName="mariadb-client-1-default" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.908334 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="76adfb82-2076-4040-a838-341d9595a9a1" containerName="mariadb-client-1-default" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.908950 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.923863 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 12 08:08:56 crc kubenswrapper[4867]: I1212 08:08:56.974258 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btb2r\" (UniqueName: \"kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r\") pod \"mariadb-client-2-default\" (UID: \"81cec280-0f77-458a-9652-1d870d09c800\") " pod="openstack/mariadb-client-2-default" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.075875 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btb2r\" (UniqueName: \"kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r\") pod \"mariadb-client-2-default\" (UID: \"81cec280-0f77-458a-9652-1d870d09c800\") " pod="openstack/mariadb-client-2-default" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.092069 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btb2r\" (UniqueName: \"kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r\") pod \"mariadb-client-2-default\" (UID: \"81cec280-0f77-458a-9652-1d870d09c800\") " pod="openstack/mariadb-client-2-default" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.162003 4867 scope.go:117] "RemoveContainer" containerID="10c7f6697285f3a68eea37a950dac79f44458cfef00879c19faa8bad0a331a98" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.162410 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.226513 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 12 08:08:57 crc kubenswrapper[4867]: I1212 08:08:57.764542 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 12 08:08:57 crc kubenswrapper[4867]: W1212 08:08:57.769117 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81cec280_0f77_458a_9652_1d870d09c800.slice/crio-401e2c0fce9718f8a66eb3b0a891bf8b7dc1936c585987532e737541a715988f WatchSource:0}: Error finding container 401e2c0fce9718f8a66eb3b0a891bf8b7dc1936c585987532e737541a715988f: Status 404 returned error can't find the container with id 401e2c0fce9718f8a66eb3b0a891bf8b7dc1936c585987532e737541a715988f Dec 12 08:08:58 crc kubenswrapper[4867]: I1212 08:08:58.175944 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"81cec280-0f77-458a-9652-1d870d09c800","Type":"ContainerStarted","Data":"a329c429f2171416dd81fafed1fdfa1d89f0de2c267689c78a956979c99ab978"} Dec 12 08:08:58 crc kubenswrapper[4867]: I1212 08:08:58.176403 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"81cec280-0f77-458a-9652-1d870d09c800","Type":"ContainerStarted","Data":"401e2c0fce9718f8a66eb3b0a891bf8b7dc1936c585987532e737541a715988f"} Dec 12 08:08:58 crc kubenswrapper[4867]: I1212 08:08:58.198473 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=2.198454049 podStartE2EDuration="2.198454049s" podCreationTimestamp="2025-12-12 08:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:08:58.197172427 +0000 UTC m=+4825.768553696" watchObservedRunningTime="2025-12-12 08:08:58.198454049 +0000 UTC m=+4825.769835318" Dec 12 08:08:59 crc kubenswrapper[4867]: I1212 08:08:59.185681 4867 generic.go:334] "Generic (PLEG): container finished" podID="81cec280-0f77-458a-9652-1d870d09c800" containerID="a329c429f2171416dd81fafed1fdfa1d89f0de2c267689c78a956979c99ab978" exitCode=1 Dec 12 08:08:59 crc kubenswrapper[4867]: I1212 08:08:59.185739 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"81cec280-0f77-458a-9652-1d870d09c800","Type":"ContainerDied","Data":"a329c429f2171416dd81fafed1fdfa1d89f0de2c267689c78a956979c99ab978"} Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.545875 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.581575 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.589478 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.729311 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btb2r\" (UniqueName: \"kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r\") pod \"81cec280-0f77-458a-9652-1d870d09c800\" (UID: \"81cec280-0f77-458a-9652-1d870d09c800\") " Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.734944 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r" (OuterVolumeSpecName: "kube-api-access-btb2r") pod "81cec280-0f77-458a-9652-1d870d09c800" (UID: "81cec280-0f77-458a-9652-1d870d09c800"). InnerVolumeSpecName "kube-api-access-btb2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.830495 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btb2r\" (UniqueName: \"kubernetes.io/projected/81cec280-0f77-458a-9652-1d870d09c800-kube-api-access-btb2r\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.855457 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81cec280-0f77-458a-9652-1d870d09c800" path="/var/lib/kubelet/pods/81cec280-0f77-458a-9652-1d870d09c800/volumes" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.958303 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 12 08:09:00 crc kubenswrapper[4867]: E1212 08:09:00.958655 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81cec280-0f77-458a-9652-1d870d09c800" containerName="mariadb-client-2-default" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.958674 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="81cec280-0f77-458a-9652-1d870d09c800" containerName="mariadb-client-2-default" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.958819 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="81cec280-0f77-458a-9652-1d870d09c800" containerName="mariadb-client-2-default" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.959309 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 12 08:09:00 crc kubenswrapper[4867]: I1212 08:09:00.971131 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.033401 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zkvj\" (UniqueName: \"kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj\") pod \"mariadb-client-1\" (UID: \"3158e644-37b1-4acb-8616-64b53f951acb\") " pod="openstack/mariadb-client-1" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.134479 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zkvj\" (UniqueName: \"kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj\") pod \"mariadb-client-1\" (UID: \"3158e644-37b1-4acb-8616-64b53f951acb\") " pod="openstack/mariadb-client-1" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.150047 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zkvj\" (UniqueName: \"kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj\") pod \"mariadb-client-1\" (UID: \"3158e644-37b1-4acb-8616-64b53f951acb\") " pod="openstack/mariadb-client-1" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.202956 4867 scope.go:117] "RemoveContainer" containerID="a329c429f2171416dd81fafed1fdfa1d89f0de2c267689c78a956979c99ab978" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.203005 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.310108 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 12 08:09:01 crc kubenswrapper[4867]: I1212 08:09:01.770018 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 12 08:09:01 crc kubenswrapper[4867]: W1212 08:09:01.773979 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3158e644_37b1_4acb_8616_64b53f951acb.slice/crio-24063c620e49020103895dfd2484d97db5bedd0c80cf720a84354f048ece181d WatchSource:0}: Error finding container 24063c620e49020103895dfd2484d97db5bedd0c80cf720a84354f048ece181d: Status 404 returned error can't find the container with id 24063c620e49020103895dfd2484d97db5bedd0c80cf720a84354f048ece181d Dec 12 08:09:02 crc kubenswrapper[4867]: I1212 08:09:02.214293 4867 generic.go:334] "Generic (PLEG): container finished" podID="3158e644-37b1-4acb-8616-64b53f951acb" containerID="22bfbc2c8546c31b207544221b6aff44ae17b65b2f1a8a60b90ab479287b0729" exitCode=0 Dec 12 08:09:02 crc kubenswrapper[4867]: I1212 08:09:02.214393 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"3158e644-37b1-4acb-8616-64b53f951acb","Type":"ContainerDied","Data":"22bfbc2c8546c31b207544221b6aff44ae17b65b2f1a8a60b90ab479287b0729"} Dec 12 08:09:02 crc kubenswrapper[4867]: I1212 08:09:02.214718 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"3158e644-37b1-4acb-8616-64b53f951acb","Type":"ContainerStarted","Data":"24063c620e49020103895dfd2484d97db5bedd0c80cf720a84354f048ece181d"} Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.552745 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.571014 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_3158e644-37b1-4acb-8616-64b53f951acb/mariadb-client-1/0.log" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.591105 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.597588 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.675477 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zkvj\" (UniqueName: \"kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj\") pod \"3158e644-37b1-4acb-8616-64b53f951acb\" (UID: \"3158e644-37b1-4acb-8616-64b53f951acb\") " Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.681422 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj" (OuterVolumeSpecName: "kube-api-access-9zkvj") pod "3158e644-37b1-4acb-8616-64b53f951acb" (UID: "3158e644-37b1-4acb-8616-64b53f951acb"). InnerVolumeSpecName "kube-api-access-9zkvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.777077 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zkvj\" (UniqueName: \"kubernetes.io/projected/3158e644-37b1-4acb-8616-64b53f951acb-kube-api-access-9zkvj\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.959033 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 12 08:09:03 crc kubenswrapper[4867]: E1212 08:09:03.959987 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3158e644-37b1-4acb-8616-64b53f951acb" containerName="mariadb-client-1" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.960012 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3158e644-37b1-4acb-8616-64b53f951acb" containerName="mariadb-client-1" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.960164 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3158e644-37b1-4acb-8616-64b53f951acb" containerName="mariadb-client-1" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.960745 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.967780 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 12 08:09:03 crc kubenswrapper[4867]: I1212 08:09:03.979282 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxzfz\" (UniqueName: \"kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz\") pod \"mariadb-client-4-default\" (UID: \"718a1c43-e32b-4ef9-a2f3-fa9764098f1c\") " pod="openstack/mariadb-client-4-default" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.080985 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxzfz\" (UniqueName: \"kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz\") pod \"mariadb-client-4-default\" (UID: \"718a1c43-e32b-4ef9-a2f3-fa9764098f1c\") " pod="openstack/mariadb-client-4-default" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.099097 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxzfz\" (UniqueName: \"kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz\") pod \"mariadb-client-4-default\" (UID: \"718a1c43-e32b-4ef9-a2f3-fa9764098f1c\") " pod="openstack/mariadb-client-4-default" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.232212 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24063c620e49020103895dfd2484d97db5bedd0c80cf720a84354f048ece181d" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.232289 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.278483 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.762928 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 12 08:09:04 crc kubenswrapper[4867]: W1212 08:09:04.768270 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718a1c43_e32b_4ef9_a2f3_fa9764098f1c.slice/crio-38ba6adbf0832d01236b48454181f44ea4a1af5f26f86bf61f2a3d84830862ba WatchSource:0}: Error finding container 38ba6adbf0832d01236b48454181f44ea4a1af5f26f86bf61f2a3d84830862ba: Status 404 returned error can't find the container with id 38ba6adbf0832d01236b48454181f44ea4a1af5f26f86bf61f2a3d84830862ba Dec 12 08:09:04 crc kubenswrapper[4867]: I1212 08:09:04.846384 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3158e644-37b1-4acb-8616-64b53f951acb" path="/var/lib/kubelet/pods/3158e644-37b1-4acb-8616-64b53f951acb/volumes" Dec 12 08:09:05 crc kubenswrapper[4867]: I1212 08:09:05.240024 4867 generic.go:334] "Generic (PLEG): container finished" podID="718a1c43-e32b-4ef9-a2f3-fa9764098f1c" containerID="6a247a5126369a743e94a2a9c7307da55fb95cec84dfe9abf3289e1cb91a9014" exitCode=0 Dec 12 08:09:05 crc kubenswrapper[4867]: I1212 08:09:05.240394 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"718a1c43-e32b-4ef9-a2f3-fa9764098f1c","Type":"ContainerDied","Data":"6a247a5126369a743e94a2a9c7307da55fb95cec84dfe9abf3289e1cb91a9014"} Dec 12 08:09:05 crc kubenswrapper[4867]: I1212 08:09:05.240513 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"718a1c43-e32b-4ef9-a2f3-fa9764098f1c","Type":"ContainerStarted","Data":"38ba6adbf0832d01236b48454181f44ea4a1af5f26f86bf61f2a3d84830862ba"} Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.601590 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.619272 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_718a1c43-e32b-4ef9-a2f3-fa9764098f1c/mariadb-client-4-default/0.log" Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.641469 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.647299 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.717706 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxzfz\" (UniqueName: \"kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz\") pod \"718a1c43-e32b-4ef9-a2f3-fa9764098f1c\" (UID: \"718a1c43-e32b-4ef9-a2f3-fa9764098f1c\") " Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.723665 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz" (OuterVolumeSpecName: "kube-api-access-nxzfz") pod "718a1c43-e32b-4ef9-a2f3-fa9764098f1c" (UID: "718a1c43-e32b-4ef9-a2f3-fa9764098f1c"). InnerVolumeSpecName "kube-api-access-nxzfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.819950 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxzfz\" (UniqueName: \"kubernetes.io/projected/718a1c43-e32b-4ef9-a2f3-fa9764098f1c-kube-api-access-nxzfz\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:06 crc kubenswrapper[4867]: I1212 08:09:06.848119 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="718a1c43-e32b-4ef9-a2f3-fa9764098f1c" path="/var/lib/kubelet/pods/718a1c43-e32b-4ef9-a2f3-fa9764098f1c/volumes" Dec 12 08:09:07 crc kubenswrapper[4867]: I1212 08:09:07.254890 4867 scope.go:117] "RemoveContainer" containerID="6a247a5126369a743e94a2a9c7307da55fb95cec84dfe9abf3289e1cb91a9014" Dec 12 08:09:07 crc kubenswrapper[4867]: I1212 08:09:07.254962 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 12 08:09:08 crc kubenswrapper[4867]: I1212 08:09:08.838614 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:09:08 crc kubenswrapper[4867]: E1212 08:09:08.839382 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.018643 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 12 08:09:10 crc kubenswrapper[4867]: E1212 08:09:10.019278 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718a1c43-e32b-4ef9-a2f3-fa9764098f1c" containerName="mariadb-client-4-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.019293 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="718a1c43-e32b-4ef9-a2f3-fa9764098f1c" containerName="mariadb-client-4-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.019445 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="718a1c43-e32b-4ef9-a2f3-fa9764098f1c" containerName="mariadb-client-4-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.019967 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.026471 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zg96v" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.037721 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.171905 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6k6l\" (UniqueName: \"kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l\") pod \"mariadb-client-5-default\" (UID: \"8f8dbd75-6b27-4a34-9233-3568fba67b36\") " pod="openstack/mariadb-client-5-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.273341 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6k6l\" (UniqueName: \"kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l\") pod \"mariadb-client-5-default\" (UID: \"8f8dbd75-6b27-4a34-9233-3568fba67b36\") " pod="openstack/mariadb-client-5-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.290822 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6k6l\" (UniqueName: \"kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l\") pod \"mariadb-client-5-default\" (UID: \"8f8dbd75-6b27-4a34-9233-3568fba67b36\") " pod="openstack/mariadb-client-5-default" Dec 12 08:09:10 crc kubenswrapper[4867]: I1212 08:09:10.340099 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 12 08:09:11 crc kubenswrapper[4867]: I1212 08:09:10.937541 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 12 08:09:11 crc kubenswrapper[4867]: W1212 08:09:10.948355 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f8dbd75_6b27_4a34_9233_3568fba67b36.slice/crio-98c121345e680471f753b200c05a55a8f6ce97b0735f6ba0b3ad7065c5dfe818 WatchSource:0}: Error finding container 98c121345e680471f753b200c05a55a8f6ce97b0735f6ba0b3ad7065c5dfe818: Status 404 returned error can't find the container with id 98c121345e680471f753b200c05a55a8f6ce97b0735f6ba0b3ad7065c5dfe818 Dec 12 08:09:11 crc kubenswrapper[4867]: I1212 08:09:11.287896 4867 generic.go:334] "Generic (PLEG): container finished" podID="8f8dbd75-6b27-4a34-9233-3568fba67b36" containerID="be136283978ef53a8576ffa31ea8cb6240748f9b42da5aa77df65b679822c5cf" exitCode=0 Dec 12 08:09:11 crc kubenswrapper[4867]: I1212 08:09:11.287951 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"8f8dbd75-6b27-4a34-9233-3568fba67b36","Type":"ContainerDied","Data":"be136283978ef53a8576ffa31ea8cb6240748f9b42da5aa77df65b679822c5cf"} Dec 12 08:09:11 crc kubenswrapper[4867]: I1212 08:09:11.288140 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"8f8dbd75-6b27-4a34-9233-3568fba67b36","Type":"ContainerStarted","Data":"98c121345e680471f753b200c05a55a8f6ce97b0735f6ba0b3ad7065c5dfe818"} Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.637180 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.653504 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_8f8dbd75-6b27-4a34-9233-3568fba67b36/mariadb-client-5-default/0.log" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.679887 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.684823 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.793448 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 12 08:09:12 crc kubenswrapper[4867]: E1212 08:09:12.793809 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8dbd75-6b27-4a34-9233-3568fba67b36" containerName="mariadb-client-5-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.793831 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8dbd75-6b27-4a34-9233-3568fba67b36" containerName="mariadb-client-5-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.794004 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8dbd75-6b27-4a34-9233-3568fba67b36" containerName="mariadb-client-5-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.794507 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.802877 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.824301 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6k6l\" (UniqueName: \"kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l\") pod \"8f8dbd75-6b27-4a34-9233-3568fba67b36\" (UID: \"8f8dbd75-6b27-4a34-9233-3568fba67b36\") " Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.830830 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l" (OuterVolumeSpecName: "kube-api-access-j6k6l") pod "8f8dbd75-6b27-4a34-9233-3568fba67b36" (UID: "8f8dbd75-6b27-4a34-9233-3568fba67b36"). InnerVolumeSpecName "kube-api-access-j6k6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.849563 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8dbd75-6b27-4a34-9233-3568fba67b36" path="/var/lib/kubelet/pods/8f8dbd75-6b27-4a34-9233-3568fba67b36/volumes" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.925925 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrms2\" (UniqueName: \"kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2\") pod \"mariadb-client-6-default\" (UID: \"056ca23b-2f8f-47c5-88d9-3df38a9db561\") " pod="openstack/mariadb-client-6-default" Dec 12 08:09:12 crc kubenswrapper[4867]: I1212 08:09:12.926082 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6k6l\" (UniqueName: \"kubernetes.io/projected/8f8dbd75-6b27-4a34-9233-3568fba67b36-kube-api-access-j6k6l\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.027708 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrms2\" (UniqueName: \"kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2\") pod \"mariadb-client-6-default\" (UID: \"056ca23b-2f8f-47c5-88d9-3df38a9db561\") " pod="openstack/mariadb-client-6-default" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.043138 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrms2\" (UniqueName: \"kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2\") pod \"mariadb-client-6-default\" (UID: \"056ca23b-2f8f-47c5-88d9-3df38a9db561\") " pod="openstack/mariadb-client-6-default" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.115156 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.303724 4867 scope.go:117] "RemoveContainer" containerID="be136283978ef53a8576ffa31ea8cb6240748f9b42da5aa77df65b679822c5cf" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.303750 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 12 08:09:13 crc kubenswrapper[4867]: I1212 08:09:13.579550 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 12 08:09:13 crc kubenswrapper[4867]: W1212 08:09:13.581371 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod056ca23b_2f8f_47c5_88d9_3df38a9db561.slice/crio-3282d6eb89dd2d4aec9560c23cd270c5461764382e4eab9ed1a6348724ed80d9 WatchSource:0}: Error finding container 3282d6eb89dd2d4aec9560c23cd270c5461764382e4eab9ed1a6348724ed80d9: Status 404 returned error can't find the container with id 3282d6eb89dd2d4aec9560c23cd270c5461764382e4eab9ed1a6348724ed80d9 Dec 12 08:09:14 crc kubenswrapper[4867]: I1212 08:09:14.312200 4867 generic.go:334] "Generic (PLEG): container finished" podID="056ca23b-2f8f-47c5-88d9-3df38a9db561" containerID="8ed629835915dc25b9d0d93c36970c9468816c30804aa7e8fa1c1c8680cde4b2" exitCode=1 Dec 12 08:09:14 crc kubenswrapper[4867]: I1212 08:09:14.312275 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"056ca23b-2f8f-47c5-88d9-3df38a9db561","Type":"ContainerDied","Data":"8ed629835915dc25b9d0d93c36970c9468816c30804aa7e8fa1c1c8680cde4b2"} Dec 12 08:09:14 crc kubenswrapper[4867]: I1212 08:09:14.312690 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"056ca23b-2f8f-47c5-88d9-3df38a9db561","Type":"ContainerStarted","Data":"3282d6eb89dd2d4aec9560c23cd270c5461764382e4eab9ed1a6348724ed80d9"} Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.679745 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.699467 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_056ca23b-2f8f-47c5-88d9-3df38a9db561/mariadb-client-6-default/0.log" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.739242 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.747455 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.869077 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrms2\" (UniqueName: \"kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2\") pod \"056ca23b-2f8f-47c5-88d9-3df38a9db561\" (UID: \"056ca23b-2f8f-47c5-88d9-3df38a9db561\") " Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.876924 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2" (OuterVolumeSpecName: "kube-api-access-xrms2") pod "056ca23b-2f8f-47c5-88d9-3df38a9db561" (UID: "056ca23b-2f8f-47c5-88d9-3df38a9db561"). InnerVolumeSpecName "kube-api-access-xrms2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.883121 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 12 08:09:15 crc kubenswrapper[4867]: E1212 08:09:15.883634 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056ca23b-2f8f-47c5-88d9-3df38a9db561" containerName="mariadb-client-6-default" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.883659 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="056ca23b-2f8f-47c5-88d9-3df38a9db561" containerName="mariadb-client-6-default" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.883811 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="056ca23b-2f8f-47c5-88d9-3df38a9db561" containerName="mariadb-client-6-default" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.884414 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.891150 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 12 08:09:15 crc kubenswrapper[4867]: I1212 08:09:15.970983 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrms2\" (UniqueName: \"kubernetes.io/projected/056ca23b-2f8f-47c5-88d9-3df38a9db561-kube-api-access-xrms2\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.072317 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9zmd\" (UniqueName: \"kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd\") pod \"mariadb-client-7-default\" (UID: \"5a43cce1-f870-45d7-b045-709f9672c111\") " pod="openstack/mariadb-client-7-default" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.173980 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9zmd\" (UniqueName: \"kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd\") pod \"mariadb-client-7-default\" (UID: \"5a43cce1-f870-45d7-b045-709f9672c111\") " pod="openstack/mariadb-client-7-default" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.190516 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9zmd\" (UniqueName: \"kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd\") pod \"mariadb-client-7-default\" (UID: \"5a43cce1-f870-45d7-b045-709f9672c111\") " pod="openstack/mariadb-client-7-default" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.222752 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.336872 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3282d6eb89dd2d4aec9560c23cd270c5461764382e4eab9ed1a6348724ed80d9" Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.336954 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 12 08:09:16 crc kubenswrapper[4867]: W1212 08:09:16.718431 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a43cce1_f870_45d7_b045_709f9672c111.slice/crio-c85fd4022439aa89549343894e24e64dfe92080f6d46a80572ab4e7d3e606c34 WatchSource:0}: Error finding container c85fd4022439aa89549343894e24e64dfe92080f6d46a80572ab4e7d3e606c34: Status 404 returned error can't find the container with id c85fd4022439aa89549343894e24e64dfe92080f6d46a80572ab4e7d3e606c34 Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.719347 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 12 08:09:16 crc kubenswrapper[4867]: I1212 08:09:16.849383 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="056ca23b-2f8f-47c5-88d9-3df38a9db561" path="/var/lib/kubelet/pods/056ca23b-2f8f-47c5-88d9-3df38a9db561/volumes" Dec 12 08:09:17 crc kubenswrapper[4867]: I1212 08:09:17.344657 4867 generic.go:334] "Generic (PLEG): container finished" podID="5a43cce1-f870-45d7-b045-709f9672c111" containerID="5c4bd4bedbbbab640401d99bd3d7588a4d29238e4a86ad4d347325bff91e13e6" exitCode=0 Dec 12 08:09:17 crc kubenswrapper[4867]: I1212 08:09:17.344909 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"5a43cce1-f870-45d7-b045-709f9672c111","Type":"ContainerDied","Data":"5c4bd4bedbbbab640401d99bd3d7588a4d29238e4a86ad4d347325bff91e13e6"} Dec 12 08:09:17 crc kubenswrapper[4867]: I1212 08:09:17.344993 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"5a43cce1-f870-45d7-b045-709f9672c111","Type":"ContainerStarted","Data":"c85fd4022439aa89549343894e24e64dfe92080f6d46a80572ab4e7d3e606c34"} Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.702674 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.728306 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9zmd\" (UniqueName: \"kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd\") pod \"5a43cce1-f870-45d7-b045-709f9672c111\" (UID: \"5a43cce1-f870-45d7-b045-709f9672c111\") " Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.730283 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_5a43cce1-f870-45d7-b045-709f9672c111/mariadb-client-7-default/0.log" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.769313 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd" (OuterVolumeSpecName: "kube-api-access-b9zmd") pod "5a43cce1-f870-45d7-b045-709f9672c111" (UID: "5a43cce1-f870-45d7-b045-709f9672c111"). InnerVolumeSpecName "kube-api-access-b9zmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.786971 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.794015 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.829757 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9zmd\" (UniqueName: \"kubernetes.io/projected/5a43cce1-f870-45d7-b045-709f9672c111-kube-api-access-b9zmd\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.846892 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a43cce1-f870-45d7-b045-709f9672c111" path="/var/lib/kubelet/pods/5a43cce1-f870-45d7-b045-709f9672c111/volumes" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.939792 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 12 08:09:18 crc kubenswrapper[4867]: E1212 08:09:18.940312 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a43cce1-f870-45d7-b045-709f9672c111" containerName="mariadb-client-7-default" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.940339 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a43cce1-f870-45d7-b045-709f9672c111" containerName="mariadb-client-7-default" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.940566 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a43cce1-f870-45d7-b045-709f9672c111" containerName="mariadb-client-7-default" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.941575 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 12 08:09:18 crc kubenswrapper[4867]: I1212 08:09:18.948987 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.032814 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k49nn\" (UniqueName: \"kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn\") pod \"mariadb-client-2\" (UID: \"cbcfd371-e48a-47fb-8314-2f832592d3c4\") " pod="openstack/mariadb-client-2" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.134875 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k49nn\" (UniqueName: \"kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn\") pod \"mariadb-client-2\" (UID: \"cbcfd371-e48a-47fb-8314-2f832592d3c4\") " pod="openstack/mariadb-client-2" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.154064 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k49nn\" (UniqueName: \"kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn\") pod \"mariadb-client-2\" (UID: \"cbcfd371-e48a-47fb-8314-2f832592d3c4\") " pod="openstack/mariadb-client-2" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.260122 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.362685 4867 scope.go:117] "RemoveContainer" containerID="5c4bd4bedbbbab640401d99bd3d7588a4d29238e4a86ad4d347325bff91e13e6" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.362857 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 12 08:09:19 crc kubenswrapper[4867]: I1212 08:09:19.549289 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 12 08:09:20 crc kubenswrapper[4867]: I1212 08:09:20.372011 4867 generic.go:334] "Generic (PLEG): container finished" podID="cbcfd371-e48a-47fb-8314-2f832592d3c4" containerID="6b93c9252fbdbe34515b9e77e48416d5b8a91f09b403c5f5f9f613bf0febc14f" exitCode=0 Dec 12 08:09:20 crc kubenswrapper[4867]: I1212 08:09:20.372099 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"cbcfd371-e48a-47fb-8314-2f832592d3c4","Type":"ContainerDied","Data":"6b93c9252fbdbe34515b9e77e48416d5b8a91f09b403c5f5f9f613bf0febc14f"} Dec 12 08:09:20 crc kubenswrapper[4867]: I1212 08:09:20.372307 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"cbcfd371-e48a-47fb-8314-2f832592d3c4","Type":"ContainerStarted","Data":"85d9bb81c08d7aa164607e2100254cec6b8bc7fd55bb097d9241ba2f6391d445"} Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.724810 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.745706 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_cbcfd371-e48a-47fb-8314-2f832592d3c4/mariadb-client-2/0.log" Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.769981 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.772272 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k49nn\" (UniqueName: \"kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn\") pod \"cbcfd371-e48a-47fb-8314-2f832592d3c4\" (UID: \"cbcfd371-e48a-47fb-8314-2f832592d3c4\") " Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.776949 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn" (OuterVolumeSpecName: "kube-api-access-k49nn") pod "cbcfd371-e48a-47fb-8314-2f832592d3c4" (UID: "cbcfd371-e48a-47fb-8314-2f832592d3c4"). InnerVolumeSpecName "kube-api-access-k49nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.777420 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.838307 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:09:21 crc kubenswrapper[4867]: E1212 08:09:21.838772 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:09:21 crc kubenswrapper[4867]: I1212 08:09:21.874811 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k49nn\" (UniqueName: \"kubernetes.io/projected/cbcfd371-e48a-47fb-8314-2f832592d3c4-kube-api-access-k49nn\") on node \"crc\" DevicePath \"\"" Dec 12 08:09:22 crc kubenswrapper[4867]: I1212 08:09:22.386534 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85d9bb81c08d7aa164607e2100254cec6b8bc7fd55bb097d9241ba2f6391d445" Dec 12 08:09:22 crc kubenswrapper[4867]: I1212 08:09:22.386611 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 12 08:09:22 crc kubenswrapper[4867]: I1212 08:09:22.848002 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbcfd371-e48a-47fb-8314-2f832592d3c4" path="/var/lib/kubelet/pods/cbcfd371-e48a-47fb-8314-2f832592d3c4/volumes" Dec 12 08:09:32 crc kubenswrapper[4867]: I1212 08:09:32.845278 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:09:33 crc kubenswrapper[4867]: I1212 08:09:33.472953 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf"} Dec 12 08:10:45 crc kubenswrapper[4867]: I1212 08:10:45.645685 4867 scope.go:117] "RemoveContainer" containerID="34a3854595e4ea075be3c5425a69508087da5185d978234c76d048c4a392abcb" Dec 12 08:11:58 crc kubenswrapper[4867]: I1212 08:11:58.988443 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:11:58 crc kubenswrapper[4867]: I1212 08:11:58.989973 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.685867 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 08:12:03 crc kubenswrapper[4867]: E1212 08:12:03.686744 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbcfd371-e48a-47fb-8314-2f832592d3c4" containerName="mariadb-client-2" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.686769 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbcfd371-e48a-47fb-8314-2f832592d3c4" containerName="mariadb-client-2" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.687011 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbcfd371-e48a-47fb-8314-2f832592d3c4" containerName="mariadb-client-2" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.688025 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.692165 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zg96v" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.703358 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.877535 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqjrq\" (UniqueName: \"kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.877648 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.978556 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqjrq\" (UniqueName: \"kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.978623 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.982238 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:03 crc kubenswrapper[4867]: I1212 08:12:03.982295 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1a2a1b1a750550e7907509bcd7396867bb06b6bfc660d7e71335d2c578bf06ed/globalmount\"" pod="openstack/mariadb-copy-data" Dec 12 08:12:04 crc kubenswrapper[4867]: I1212 08:12:04.000348 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqjrq\" (UniqueName: \"kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:04 crc kubenswrapper[4867]: I1212 08:12:04.009326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") pod \"mariadb-copy-data\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " pod="openstack/mariadb-copy-data" Dec 12 08:12:04 crc kubenswrapper[4867]: I1212 08:12:04.307089 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 12 08:12:04 crc kubenswrapper[4867]: W1212 08:12:04.843579 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76e39b07_673e_40d8_a2ec_3f6fbd894418.slice/crio-5026f6b8f633c9ee142eb0855d4136cfe5ee6612a218e18ec8d78a1d932e906e WatchSource:0}: Error finding container 5026f6b8f633c9ee142eb0855d4136cfe5ee6612a218e18ec8d78a1d932e906e: Status 404 returned error can't find the container with id 5026f6b8f633c9ee142eb0855d4136cfe5ee6612a218e18ec8d78a1d932e906e Dec 12 08:12:04 crc kubenswrapper[4867]: I1212 08:12:04.848100 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 08:12:04 crc kubenswrapper[4867]: I1212 08:12:04.913611 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"76e39b07-673e-40d8-a2ec-3f6fbd894418","Type":"ContainerStarted","Data":"5026f6b8f633c9ee142eb0855d4136cfe5ee6612a218e18ec8d78a1d932e906e"} Dec 12 08:12:05 crc kubenswrapper[4867]: I1212 08:12:05.922996 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"76e39b07-673e-40d8-a2ec-3f6fbd894418","Type":"ContainerStarted","Data":"9dadee254a52c43da417caeeef0783d37846dea1fe84f9b21c4541f7826403df"} Dec 12 08:12:05 crc kubenswrapper[4867]: I1212 08:12:05.943074 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.943054075 podStartE2EDuration="3.943054075s" podCreationTimestamp="2025-12-12 08:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:12:05.936040203 +0000 UTC m=+5013.507421482" watchObservedRunningTime="2025-12-12 08:12:05.943054075 +0000 UTC m=+5013.514435344" Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.583896 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.585921 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.592097 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.749453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svg74\" (UniqueName: \"kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74\") pod \"mariadb-client\" (UID: \"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8\") " pod="openstack/mariadb-client" Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.850873 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svg74\" (UniqueName: \"kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74\") pod \"mariadb-client\" (UID: \"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8\") " pod="openstack/mariadb-client" Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.870140 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svg74\" (UniqueName: \"kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74\") pod \"mariadb-client\" (UID: \"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8\") " pod="openstack/mariadb-client" Dec 12 08:12:08 crc kubenswrapper[4867]: I1212 08:12:08.912303 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:09 crc kubenswrapper[4867]: I1212 08:12:09.328523 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:09 crc kubenswrapper[4867]: I1212 08:12:09.952471 4867 generic.go:334] "Generic (PLEG): container finished" podID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" containerID="113dd68440afbde4000b8c6e3b5f4297d39cce26c25685759fd310153d2d57a4" exitCode=0 Dec 12 08:12:09 crc kubenswrapper[4867]: I1212 08:12:09.952532 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8","Type":"ContainerDied","Data":"113dd68440afbde4000b8c6e3b5f4297d39cce26c25685759fd310153d2d57a4"} Dec 12 08:12:09 crc kubenswrapper[4867]: I1212 08:12:09.952768 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8","Type":"ContainerStarted","Data":"bdfb7037e0ea2f30e1d1f6a41665051c025e4bfd10dfca15f77ae99b2af4860d"} Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.284219 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.309145 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6e1385fb-d7fe-4849-9ef4-7ce55113c5b8/mariadb-client/0.log" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.340988 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.349629 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.388461 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svg74\" (UniqueName: \"kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74\") pod \"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8\" (UID: \"6e1385fb-d7fe-4849-9ef4-7ce55113c5b8\") " Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.395992 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74" (OuterVolumeSpecName: "kube-api-access-svg74") pod "6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" (UID: "6e1385fb-d7fe-4849-9ef4-7ce55113c5b8"). InnerVolumeSpecName "kube-api-access-svg74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.462010 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:11 crc kubenswrapper[4867]: E1212 08:12:11.462563 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" containerName="mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.462590 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" containerName="mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.463024 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" containerName="mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.463822 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.469798 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.490324 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svg74\" (UniqueName: \"kubernetes.io/projected/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8-kube-api-access-svg74\") on node \"crc\" DevicePath \"\"" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.591487 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79ljn\" (UniqueName: \"kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn\") pod \"mariadb-client\" (UID: \"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7\") " pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.692670 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79ljn\" (UniqueName: \"kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn\") pod \"mariadb-client\" (UID: \"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7\") " pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.710427 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79ljn\" (UniqueName: \"kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn\") pod \"mariadb-client\" (UID: \"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7\") " pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.784634 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.969550 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdfb7037e0ea2f30e1d1f6a41665051c025e4bfd10dfca15f77ae99b2af4860d" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.970068 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:11 crc kubenswrapper[4867]: I1212 08:12:11.989889 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" podUID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" Dec 12 08:12:12 crc kubenswrapper[4867]: I1212 08:12:12.216125 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:12 crc kubenswrapper[4867]: W1212 08:12:12.220560 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3168d1ce_a2b4_4ff3_b1e9_9157c36e2cc7.slice/crio-d9bfce46b40169fa9762a094dce9b107597b71214f206916784fee9735e69080 WatchSource:0}: Error finding container d9bfce46b40169fa9762a094dce9b107597b71214f206916784fee9735e69080: Status 404 returned error can't find the container with id d9bfce46b40169fa9762a094dce9b107597b71214f206916784fee9735e69080 Dec 12 08:12:12 crc kubenswrapper[4867]: I1212 08:12:12.874453 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e1385fb-d7fe-4849-9ef4-7ce55113c5b8" path="/var/lib/kubelet/pods/6e1385fb-d7fe-4849-9ef4-7ce55113c5b8/volumes" Dec 12 08:12:12 crc kubenswrapper[4867]: I1212 08:12:12.979659 4867 generic.go:334] "Generic (PLEG): container finished" podID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" containerID="2b1dddecccd4e5ea07d851c10d4f3c2857d9f74515d521d1b368397a0386aa02" exitCode=0 Dec 12 08:12:12 crc kubenswrapper[4867]: I1212 08:12:12.979734 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7","Type":"ContainerDied","Data":"2b1dddecccd4e5ea07d851c10d4f3c2857d9f74515d521d1b368397a0386aa02"} Dec 12 08:12:12 crc kubenswrapper[4867]: I1212 08:12:12.979780 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7","Type":"ContainerStarted","Data":"d9bfce46b40169fa9762a094dce9b107597b71214f206916784fee9735e69080"} Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.287329 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.303986 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7/mariadb-client/0.log" Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.330443 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.337443 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.432209 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79ljn\" (UniqueName: \"kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn\") pod \"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7\" (UID: \"3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7\") " Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.437587 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn" (OuterVolumeSpecName: "kube-api-access-79ljn") pod "3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" (UID: "3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7"). InnerVolumeSpecName "kube-api-access-79ljn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.534200 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79ljn\" (UniqueName: \"kubernetes.io/projected/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7-kube-api-access-79ljn\") on node \"crc\" DevicePath \"\"" Dec 12 08:12:14 crc kubenswrapper[4867]: I1212 08:12:14.846746 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" path="/var/lib/kubelet/pods/3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7/volumes" Dec 12 08:12:15 crc kubenswrapper[4867]: I1212 08:12:15.002306 4867 scope.go:117] "RemoveContainer" containerID="2b1dddecccd4e5ea07d851c10d4f3c2857d9f74515d521d1b368397a0386aa02" Dec 12 08:12:15 crc kubenswrapper[4867]: I1212 08:12:15.002340 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 12 08:12:28 crc kubenswrapper[4867]: I1212 08:12:28.989646 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:12:28 crc kubenswrapper[4867]: I1212 08:12:28.990319 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.809028 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 08:12:49 crc kubenswrapper[4867]: E1212 08:12:49.810090 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" containerName="mariadb-client" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.810109 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" containerName="mariadb-client" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.810461 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3168d1ce-a2b4-4ff3-b1e9-9157c36e2cc7" containerName="mariadb-client" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.811857 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.813713 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-llh7d" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.814167 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.814516 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.814651 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.814936 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.816792 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.818248 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.830491 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.833591 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.862364 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.870921 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.883550 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964598 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nsft\" (UniqueName: \"kubernetes.io/projected/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-kube-api-access-5nsft\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964681 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964705 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964768 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964843 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964877 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964907 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzd8\" (UniqueName: \"kubernetes.io/projected/709a4369-8271-49da-b10e-810db17bb0db-kube-api-access-hpzd8\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964932 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964970 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.964997 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965020 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-76611231-e53b-4308-a127-2de4b8d9862e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-76611231-e53b-4308-a127-2de4b8d9862e\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965142 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965207 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965376 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965410 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965478 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965532 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-config\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965553 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/709a4369-8271-49da-b10e-810db17bb0db-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965597 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-config\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965635 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-config\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965658 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncdk6\" (UniqueName: \"kubernetes.io/projected/f852bd2b-64b9-432a-a921-1f16c6f67cea-kube-api-access-ncdk6\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965689 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:49 crc kubenswrapper[4867]: I1212 08:12:49.965712 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066727 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066774 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066808 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066833 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-config\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066848 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/709a4369-8271-49da-b10e-810db17bb0db-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066873 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-config\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-config\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066935 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncdk6\" (UniqueName: \"kubernetes.io/projected/f852bd2b-64b9-432a-a921-1f16c6f67cea-kube-api-access-ncdk6\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066965 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.066989 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067013 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nsft\" (UniqueName: \"kubernetes.io/projected/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-kube-api-access-5nsft\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067044 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067068 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067091 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067114 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067145 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067167 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067198 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzd8\" (UniqueName: \"kubernetes.io/projected/709a4369-8271-49da-b10e-810db17bb0db-kube-api-access-hpzd8\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067241 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067281 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067308 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067333 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-76611231-e53b-4308-a127-2de4b8d9862e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-76611231-e53b-4308-a127-2de4b8d9862e\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067373 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067404 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067585 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/709a4369-8271-49da-b10e-810db17bb0db-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.067976 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-config\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.068036 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-config\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.068384 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.068402 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.069414 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-config\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.069550 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/709a4369-8271-49da-b10e-810db17bb0db-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.069632 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f852bd2b-64b9-432a-a921-1f16c6f67cea-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.069857 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.073763 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.073802 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-76611231-e53b-4308-a127-2de4b8d9862e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-76611231-e53b-4308-a127-2de4b8d9862e\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b464516c6b89eb12433d7bcc9c03ea6548253e5f5d4a7150f51e0c50eeadb1ea/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.073970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.074014 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.074029 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.074489 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.074519 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/940cee92c445967d50e2584ee688a1daab08cc280a40f8d5139f98690aceee0b/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.075497 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.075798 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.075997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.076965 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f852bd2b-64b9-432a-a921-1f16c6f67cea-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.077813 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/709a4369-8271-49da-b10e-810db17bb0db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.081256 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.084098 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncdk6\" (UniqueName: \"kubernetes.io/projected/f852bd2b-64b9-432a-a921-1f16c6f67cea-kube-api-access-ncdk6\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.084797 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzd8\" (UniqueName: \"kubernetes.io/projected/709a4369-8271-49da-b10e-810db17bb0db-kube-api-access-hpzd8\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.088894 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.088925 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9f58dfc09ae6ae95e22164d6d7bd82e9dc2e0c036a3cd92ceb4d4952c652de3a/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.092657 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nsft\" (UniqueName: \"kubernetes.io/projected/dd670fea-d4c3-4a38-91c5-918f6db4fb5b-kube-api-access-5nsft\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.106681 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-15c6b5f4-2ff2-4c80-8433-d6571824cd33\") pod \"ovsdbserver-nb-1\" (UID: \"709a4369-8271-49da-b10e-810db17bb0db\") " pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.108382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-76611231-e53b-4308-a127-2de4b8d9862e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-76611231-e53b-4308-a127-2de4b8d9862e\") pod \"ovsdbserver-nb-0\" (UID: \"f852bd2b-64b9-432a-a921-1f16c6f67cea\") " pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.119805 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6cec008-7b03-4a74-b928-62b406d3dc2a\") pod \"ovsdbserver-nb-2\" (UID: \"dd670fea-d4c3-4a38-91c5-918f6db4fb5b\") " pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.141088 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.153905 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.164278 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.739187 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 12 08:12:50 crc kubenswrapper[4867]: I1212 08:12:50.799091 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 12 08:12:50 crc kubenswrapper[4867]: W1212 08:12:50.800653 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod709a4369_8271_49da_b10e_810db17bb0db.slice/crio-5a5c39d9e1f8918150285002e6e11bf41070a61a7b319066deb707c2f684198e WatchSource:0}: Error finding container 5a5c39d9e1f8918150285002e6e11bf41070a61a7b319066deb707c2f684198e: Status 404 returned error can't find the container with id 5a5c39d9e1f8918150285002e6e11bf41070a61a7b319066deb707c2f684198e Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.277082 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"709a4369-8271-49da-b10e-810db17bb0db","Type":"ContainerStarted","Data":"5a5c39d9e1f8918150285002e6e11bf41070a61a7b319066deb707c2f684198e"} Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.278188 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f852bd2b-64b9-432a-a921-1f16c6f67cea","Type":"ContainerStarted","Data":"7174d13ee932a94a482415c9fc1abbc030cf09d4fa77135481078004e9578fc4"} Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.594567 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.893341 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.897563 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.900260 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.900409 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.900423 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-gn9vb" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.900626 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.903351 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.949261 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.951074 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.976353 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.977805 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:51 crc kubenswrapper[4867]: I1212 08:12:51.995103 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.003931 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010169 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010270 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010443 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-config\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010536 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010601 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010639 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010682 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.010737 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx5ts\" (UniqueName: \"kubernetes.io/projected/36c35414-d05a-45e6-92c0-f28e80b01c00-kube-api-access-vx5ts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112031 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-config\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112087 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-config\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112107 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112127 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtxds\" (UniqueName: \"kubernetes.io/projected/1ddac016-beca-4499-9ac8-2b846fb51a03-kube-api-access-mtxds\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112152 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112178 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112201 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112219 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112254 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr97r\" (UniqueName: \"kubernetes.io/projected/f98a91b8-7475-452a-a7fe-eafbd824c99f-kube-api-access-zr97r\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112271 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112290 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112307 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx5ts\" (UniqueName: \"kubernetes.io/projected/36c35414-d05a-45e6-92c0-f28e80b01c00-kube-api-access-vx5ts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112336 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112354 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112378 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112397 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-config\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112419 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112454 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112475 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112495 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112512 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112531 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112550 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-01bb6961-4100-46bf-9721-e6af909444b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-01bb6961-4100-46bf-9721-e6af909444b4\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.112567 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.113548 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-config\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.113956 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.117096 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36c35414-d05a-45e6-92c0-f28e80b01c00-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.120577 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.121401 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.123615 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.123671 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fc99e49a9d1b47f97a4e6a412b154b5fab7538d24bddda8a9ca228b8f1b4cd48/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.131784 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c35414-d05a-45e6-92c0-f28e80b01c00-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.133092 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx5ts\" (UniqueName: \"kubernetes.io/projected/36c35414-d05a-45e6-92c0-f28e80b01c00-kube-api-access-vx5ts\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.150475 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8f8737d6-810b-4373-9230-94ec8cd12cb6\") pod \"ovsdbserver-sb-0\" (UID: \"36c35414-d05a-45e6-92c0-f28e80b01c00\") " pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214026 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-01bb6961-4100-46bf-9721-e6af909444b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-01bb6961-4100-46bf-9721-e6af909444b4\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214082 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214122 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-config\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214142 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214161 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtxds\" (UniqueName: \"kubernetes.io/projected/1ddac016-beca-4499-9ac8-2b846fb51a03-kube-api-access-mtxds\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214195 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214244 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr97r\" (UniqueName: \"kubernetes.io/projected/f98a91b8-7475-452a-a7fe-eafbd824c99f-kube-api-access-zr97r\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214275 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214317 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214339 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214369 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-config\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214411 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214452 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214483 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214501 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.214989 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.215527 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.215885 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.216462 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f98a91b8-7475-452a-a7fe-eafbd824c99f-config\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.216626 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.217755 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ddac016-beca-4499-9ac8-2b846fb51a03-config\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.219141 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.224464 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.224512 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ce6188bb4d577f815ffe2ba1033fba5b4dfc75020029d09acfec18f4bdef90b0/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.224828 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.224925 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-01bb6961-4100-46bf-9721-e6af909444b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-01bb6961-4100-46bf-9721-e6af909444b4\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2877791df64380318962db5f7af853bf6e42c1751f8c1dbd1edb04b23a9ffb2e/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.225436 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.227315 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.227428 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.228058 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f98a91b8-7475-452a-a7fe-eafbd824c99f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.232969 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr97r\" (UniqueName: \"kubernetes.io/projected/f98a91b8-7475-452a-a7fe-eafbd824c99f-kube-api-access-zr97r\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.233171 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddac016-beca-4499-9ac8-2b846fb51a03-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.235126 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtxds\" (UniqueName: \"kubernetes.io/projected/1ddac016-beca-4499-9ac8-2b846fb51a03-kube-api-access-mtxds\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.245247 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.261070 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-01bb6961-4100-46bf-9721-e6af909444b4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-01bb6961-4100-46bf-9721-e6af909444b4\") pod \"ovsdbserver-sb-1\" (UID: \"1ddac016-beca-4499-9ac8-2b846fb51a03\") " pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.264222 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9e92dae7-035a-4b18-964a-b8c8b02f8ade\") pod \"ovsdbserver-sb-2\" (UID: \"f98a91b8-7475-452a-a7fe-eafbd824c99f\") " pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.275978 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.296563 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.297291 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"dd670fea-d4c3-4a38-91c5-918f6db4fb5b","Type":"ContainerStarted","Data":"3ccea251a780f46c90dfa72048b7fd4a73457ac4b8b867e61b999cca8fde11ae"} Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.795031 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 12 08:12:52 crc kubenswrapper[4867]: W1212 08:12:52.802980 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36c35414_d05a_45e6_92c0_f28e80b01c00.slice/crio-65393c6ddbb2568df2456aacdfc1a647fc6c0a212a7f4574ca52ba6e3faa58ab WatchSource:0}: Error finding container 65393c6ddbb2568df2456aacdfc1a647fc6c0a212a7f4574ca52ba6e3faa58ab: Status 404 returned error can't find the container with id 65393c6ddbb2568df2456aacdfc1a647fc6c0a212a7f4574ca52ba6e3faa58ab Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.933051 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 12 08:12:52 crc kubenswrapper[4867]: W1212 08:12:52.935270 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ddac016_beca_4499_9ac8_2b846fb51a03.slice/crio-fd0b6cfa978adddd33972c9af24d335078d53c93734f7c826ea779cda5c5104f WatchSource:0}: Error finding container fd0b6cfa978adddd33972c9af24d335078d53c93734f7c826ea779cda5c5104f: Status 404 returned error can't find the container with id fd0b6cfa978adddd33972c9af24d335078d53c93734f7c826ea779cda5c5104f Dec 12 08:12:52 crc kubenswrapper[4867]: I1212 08:12:52.997255 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 12 08:12:53 crc kubenswrapper[4867]: W1212 08:12:53.006382 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf98a91b8_7475_452a_a7fe_eafbd824c99f.slice/crio-520db6f1154a58b23d13f4b10865dfa4deb388e576e749b4dac8c464334a6246 WatchSource:0}: Error finding container 520db6f1154a58b23d13f4b10865dfa4deb388e576e749b4dac8c464334a6246: Status 404 returned error can't find the container with id 520db6f1154a58b23d13f4b10865dfa4deb388e576e749b4dac8c464334a6246 Dec 12 08:12:53 crc kubenswrapper[4867]: I1212 08:12:53.309653 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f98a91b8-7475-452a-a7fe-eafbd824c99f","Type":"ContainerStarted","Data":"520db6f1154a58b23d13f4b10865dfa4deb388e576e749b4dac8c464334a6246"} Dec 12 08:12:53 crc kubenswrapper[4867]: I1212 08:12:53.311939 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36c35414-d05a-45e6-92c0-f28e80b01c00","Type":"ContainerStarted","Data":"65393c6ddbb2568df2456aacdfc1a647fc6c0a212a7f4574ca52ba6e3faa58ab"} Dec 12 08:12:53 crc kubenswrapper[4867]: I1212 08:12:53.313145 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"1ddac016-beca-4499-9ac8-2b846fb51a03","Type":"ContainerStarted","Data":"fd0b6cfa978adddd33972c9af24d335078d53c93734f7c826ea779cda5c5104f"} Dec 12 08:12:55 crc kubenswrapper[4867]: I1212 08:12:55.333315 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"dd670fea-d4c3-4a38-91c5-918f6db4fb5b","Type":"ContainerStarted","Data":"ddd30d2672d460e92c58f2ca6edfb1b6e37f59623cbcfbf8e5717cc6e97b1c56"} Dec 12 08:12:55 crc kubenswrapper[4867]: I1212 08:12:55.334861 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"709a4369-8271-49da-b10e-810db17bb0db","Type":"ContainerStarted","Data":"f3d41a0f6487eaee26148ed022204862144b2dc4c0d5a73368fdb3220e1825a3"} Dec 12 08:12:55 crc kubenswrapper[4867]: I1212 08:12:55.336575 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f852bd2b-64b9-432a-a921-1f16c6f67cea","Type":"ContainerStarted","Data":"cb17d57d948665fdb5b85371b35aaa9e76e885cb8661aa0d561bb285bd5b6998"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.346239 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"dd670fea-d4c3-4a38-91c5-918f6db4fb5b","Type":"ContainerStarted","Data":"a486939a4988cafc274c50f37c8be0ccbf3bdc00339fce6976cb330ad60cf4eb"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.348173 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36c35414-d05a-45e6-92c0-f28e80b01c00","Type":"ContainerStarted","Data":"db460ed7fcfc86b7d763203b23331ba07644ca59a99187a9a85817f2f0288fc7"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.348500 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36c35414-d05a-45e6-92c0-f28e80b01c00","Type":"ContainerStarted","Data":"5d0e5fe32d269b44f80b6b5f0d467474362cd2c8dacd6d25a076448956cabbf8"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.350555 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"709a4369-8271-49da-b10e-810db17bb0db","Type":"ContainerStarted","Data":"a23c81fbfddb0fc0f92b13cf559300a43c6073012895ca4e4aabce8e1f9483e2"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.352415 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"1ddac016-beca-4499-9ac8-2b846fb51a03","Type":"ContainerStarted","Data":"cc1b7482aa4ea333cad1c32c4a7fbccb4d8db9ec653c751304bc67158a9da351"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.352525 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"1ddac016-beca-4499-9ac8-2b846fb51a03","Type":"ContainerStarted","Data":"ab99275e5cb2b31bf86466170e08c504f9737643dd3b406cd40cccf4f57c5646"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.354081 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f98a91b8-7475-452a-a7fe-eafbd824c99f","Type":"ContainerStarted","Data":"36baefad811f79c671b417f585fcfe520da99d0416107fbc6528b1182f0dcd2f"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.354117 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"f98a91b8-7475-452a-a7fe-eafbd824c99f","Type":"ContainerStarted","Data":"af5d41e6c66ac3b75a1281206779fa820e96e664aee4df3ab54a6ae344480fb7"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.355578 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f852bd2b-64b9-432a-a921-1f16c6f67cea","Type":"ContainerStarted","Data":"ac3058c5a9295f4bced4a6095e428c5ae04205afb3be6a202400f06697e94cb0"} Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.374400 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=5.071923835 podStartE2EDuration="8.374383626s" podCreationTimestamp="2025-12-12 08:12:48 +0000 UTC" firstStartedPulling="2025-12-12 08:12:51.606855435 +0000 UTC m=+5059.178236704" lastFinishedPulling="2025-12-12 08:12:54.909315226 +0000 UTC m=+5062.480696495" observedRunningTime="2025-12-12 08:12:56.371939506 +0000 UTC m=+5063.943320775" watchObservedRunningTime="2025-12-12 08:12:56.374383626 +0000 UTC m=+5063.945764895" Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.393480 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.834509691 podStartE2EDuration="6.393463556s" podCreationTimestamp="2025-12-12 08:12:50 +0000 UTC" firstStartedPulling="2025-12-12 08:12:52.937219737 +0000 UTC m=+5060.508601006" lastFinishedPulling="2025-12-12 08:12:55.496173602 +0000 UTC m=+5063.067554871" observedRunningTime="2025-12-12 08:12:56.389687723 +0000 UTC m=+5063.961068992" watchObservedRunningTime="2025-12-12 08:12:56.393463556 +0000 UTC m=+5063.964844825" Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.416005 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.9294629690000002 podStartE2EDuration="6.41596947s" podCreationTimestamp="2025-12-12 08:12:50 +0000 UTC" firstStartedPulling="2025-12-12 08:12:53.008847552 +0000 UTC m=+5060.580228831" lastFinishedPulling="2025-12-12 08:12:55.495354063 +0000 UTC m=+5063.066735332" observedRunningTime="2025-12-12 08:12:56.408456875 +0000 UTC m=+5063.979838144" watchObservedRunningTime="2025-12-12 08:12:56.41596947 +0000 UTC m=+5063.987350739" Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.432286 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.741185462 podStartE2EDuration="6.432255412s" podCreationTimestamp="2025-12-12 08:12:50 +0000 UTC" firstStartedPulling="2025-12-12 08:12:52.805522793 +0000 UTC m=+5060.376904062" lastFinishedPulling="2025-12-12 08:12:55.496592743 +0000 UTC m=+5063.067974012" observedRunningTime="2025-12-12 08:12:56.424466649 +0000 UTC m=+5063.995847938" watchObservedRunningTime="2025-12-12 08:12:56.432255412 +0000 UTC m=+5064.003636681" Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.447262 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.366448338 podStartE2EDuration="8.447216561s" podCreationTimestamp="2025-12-12 08:12:48 +0000 UTC" firstStartedPulling="2025-12-12 08:12:50.804175563 +0000 UTC m=+5058.375556832" lastFinishedPulling="2025-12-12 08:12:54.884943786 +0000 UTC m=+5062.456325055" observedRunningTime="2025-12-12 08:12:56.441445288 +0000 UTC m=+5064.012826567" watchObservedRunningTime="2025-12-12 08:12:56.447216561 +0000 UTC m=+5064.018597830" Dec 12 08:12:56 crc kubenswrapper[4867]: I1212 08:12:56.472264 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.297678612 podStartE2EDuration="8.472243496s" podCreationTimestamp="2025-12-12 08:12:48 +0000 UTC" firstStartedPulling="2025-12-12 08:12:50.748255855 +0000 UTC m=+5058.319637124" lastFinishedPulling="2025-12-12 08:12:54.922820739 +0000 UTC m=+5062.494202008" observedRunningTime="2025-12-12 08:12:56.468116215 +0000 UTC m=+5064.039497484" watchObservedRunningTime="2025-12-12 08:12:56.472243496 +0000 UTC m=+5064.043624765" Dec 12 08:12:57 crc kubenswrapper[4867]: I1212 08:12:57.245915 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:57 crc kubenswrapper[4867]: I1212 08:12:57.276427 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:57 crc kubenswrapper[4867]: I1212 08:12:57.296820 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.246345 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.276414 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.284313 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.297863 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.322837 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.333319 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.989088 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.989146 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.989186 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.989898 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:12:58 crc kubenswrapper[4867]: I1212 08:12:58.989965 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf" gracePeriod=600 Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.142109 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.155179 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.164676 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.188736 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.197813 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.216384 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.381008 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf" exitCode=0 Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.381087 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf"} Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.382383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570"} Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.382419 4867 scope.go:117] "RemoveContainer" containerID="b076628150d38422b81b5ae8630226cdf664cbad3dcad18e213bded53acfc44f" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.382558 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.384463 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 12 08:12:59 crc kubenswrapper[4867]: I1212 08:12:59.384515 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.185432 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.190688 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.212642 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.422879 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.434155 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.440264 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.441418 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.557900 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.560521 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.560649 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrrj\" (UniqueName: \"kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.560687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.662922 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.663000 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.663042 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrrj\" (UniqueName: \"kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.663064 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.664094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.664144 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.664834 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.687012 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrrj\" (UniqueName: \"kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj\") pod \"dnsmasq-dns-8fdb65495-jfxkp\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:00 crc kubenswrapper[4867]: I1212 08:13:00.761174 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:01 crc kubenswrapper[4867]: I1212 08:13:01.217958 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:01 crc kubenswrapper[4867]: W1212 08:13:01.227396 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b343d7_d7cb_4376_be0d_d2f6d229959f.slice/crio-567bb7e965fa70b50262b06adc6ec7d626fe215e4132a2aa07e03a457b4feddb WatchSource:0}: Error finding container 567bb7e965fa70b50262b06adc6ec7d626fe215e4132a2aa07e03a457b4feddb: Status 404 returned error can't find the container with id 567bb7e965fa70b50262b06adc6ec7d626fe215e4132a2aa07e03a457b4feddb Dec 12 08:13:01 crc kubenswrapper[4867]: I1212 08:13:01.401299 4867 generic.go:334] "Generic (PLEG): container finished" podID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerID="0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9" exitCode=0 Dec 12 08:13:01 crc kubenswrapper[4867]: I1212 08:13:01.401435 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" event={"ID":"f5b343d7-d7cb-4376-be0d-d2f6d229959f","Type":"ContainerDied","Data":"0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9"} Dec 12 08:13:01 crc kubenswrapper[4867]: I1212 08:13:01.401475 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" event={"ID":"f5b343d7-d7cb-4376-be0d-d2f6d229959f","Type":"ContainerStarted","Data":"567bb7e965fa70b50262b06adc6ec7d626fe215e4132a2aa07e03a457b4feddb"} Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.285184 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.321607 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.358309 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.423844 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" event={"ID":"f5b343d7-d7cb-4376-be0d-d2f6d229959f","Type":"ContainerStarted","Data":"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261"} Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.424031 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.445445 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" podStartSLOduration=2.445420327 podStartE2EDuration="2.445420327s" podCreationTimestamp="2025-12-12 08:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:02.441677654 +0000 UTC m=+5070.013058933" watchObservedRunningTime="2025-12-12 08:13:02.445420327 +0000 UTC m=+5070.016801596" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.502723 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.549470 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.550832 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.553100 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.562584 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.594934 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.594990 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzz6q\" (UniqueName: \"kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.595111 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.595153 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.595170 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.696459 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.696516 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzz6q\" (UniqueName: \"kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.696578 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.696610 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.696629 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.697624 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.697695 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.697701 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.697947 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.715620 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzz6q\" (UniqueName: \"kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q\") pod \"dnsmasq-dns-6cd56d8775-cfgzb\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:02 crc kubenswrapper[4867]: I1212 08:13:02.874799 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:03 crc kubenswrapper[4867]: I1212 08:13:03.325897 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:03 crc kubenswrapper[4867]: W1212 08:13:03.330417 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1e6b2fd_8b09_4498_9fbe_c1b27625f264.slice/crio-9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c WatchSource:0}: Error finding container 9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c: Status 404 returned error can't find the container with id 9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c Dec 12 08:13:03 crc kubenswrapper[4867]: I1212 08:13:03.431259 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" event={"ID":"d1e6b2fd-8b09-4498-9fbe-c1b27625f264","Type":"ContainerStarted","Data":"9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c"} Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.440254 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerID="87a8dd351b2ace9a29ffd52442b4dce7dff5ec54d3828f0fbaf4f2be1fdaaf0b" exitCode=0 Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.440335 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" event={"ID":"d1e6b2fd-8b09-4498-9fbe-c1b27625f264","Type":"ContainerDied","Data":"87a8dd351b2ace9a29ffd52442b4dce7dff5ec54d3828f0fbaf4f2be1fdaaf0b"} Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.440703 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="dnsmasq-dns" containerID="cri-o://d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261" gracePeriod=10 Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.725515 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.730350 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.732579 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.735186 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.829551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.829636 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v94gg\" (UniqueName: \"kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.829836 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.884529 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.931616 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config\") pod \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.931777 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb\") pod \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.931847 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc\") pod \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.932562 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmrrj\" (UniqueName: \"kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj\") pod \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\" (UID: \"f5b343d7-d7cb-4376-be0d-d2f6d229959f\") " Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.933336 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.933452 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.933582 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v94gg\" (UniqueName: \"kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.941182 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.941570 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.941660 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bebaef9c44e1c0c5200c502a25385628b6d49f9f056bb58ca11a443fd722d3b7/globalmount\"" pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.942828 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj" (OuterVolumeSpecName: "kube-api-access-fmrrj") pod "f5b343d7-d7cb-4376-be0d-d2f6d229959f" (UID: "f5b343d7-d7cb-4376-be0d-d2f6d229959f"). InnerVolumeSpecName "kube-api-access-fmrrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.957518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v94gg\" (UniqueName: \"kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.977287 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") pod \"ovn-copy-data\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " pod="openstack/ovn-copy-data" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.979568 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5b343d7-d7cb-4376-be0d-d2f6d229959f" (UID: "f5b343d7-d7cb-4376-be0d-d2f6d229959f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.982159 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config" (OuterVolumeSpecName: "config") pod "f5b343d7-d7cb-4376-be0d-d2f6d229959f" (UID: "f5b343d7-d7cb-4376-be0d-d2f6d229959f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:04 crc kubenswrapper[4867]: I1212 08:13:04.985245 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5b343d7-d7cb-4376-be0d-d2f6d229959f" (UID: "f5b343d7-d7cb-4376-be0d-d2f6d229959f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.034406 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.034433 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.034443 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5b343d7-d7cb-4376-be0d-d2f6d229959f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.034451 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmrrj\" (UniqueName: \"kubernetes.io/projected/f5b343d7-d7cb-4376-be0d-d2f6d229959f-kube-api-access-fmrrj\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.058015 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.454390 4867 generic.go:334] "Generic (PLEG): container finished" podID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerID="d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261" exitCode=0 Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.454768 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" event={"ID":"f5b343d7-d7cb-4376-be0d-d2f6d229959f","Type":"ContainerDied","Data":"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261"} Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.454951 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" event={"ID":"f5b343d7-d7cb-4376-be0d-d2f6d229959f","Type":"ContainerDied","Data":"567bb7e965fa70b50262b06adc6ec7d626fe215e4132a2aa07e03a457b4feddb"} Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.454996 4867 scope.go:117] "RemoveContainer" containerID="d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.455067 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fdb65495-jfxkp" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.461127 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" event={"ID":"d1e6b2fd-8b09-4498-9fbe-c1b27625f264","Type":"ContainerStarted","Data":"7525e1e93a72e8fb116d175998e511e0b9a6955a80e82c3df82426e39d1e7c4e"} Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.461413 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.483863 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" podStartSLOduration=3.483778362 podStartE2EDuration="3.483778362s" podCreationTimestamp="2025-12-12 08:13:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:05.482733206 +0000 UTC m=+5073.054114475" watchObservedRunningTime="2025-12-12 08:13:05.483778362 +0000 UTC m=+5073.055159621" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.487851 4867 scope.go:117] "RemoveContainer" containerID="0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.509270 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.510411 4867 scope.go:117] "RemoveContainer" containerID="d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261" Dec 12 08:13:05 crc kubenswrapper[4867]: E1212 08:13:05.510887 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261\": container with ID starting with d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261 not found: ID does not exist" containerID="d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.510928 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261"} err="failed to get container status \"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261\": rpc error: code = NotFound desc = could not find container \"d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261\": container with ID starting with d9f72c3d5c6cc1e4c7af38a91c80e1433ae2def48c378d6676313b179defe261 not found: ID does not exist" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.510955 4867 scope.go:117] "RemoveContainer" containerID="0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9" Dec 12 08:13:05 crc kubenswrapper[4867]: E1212 08:13:05.511334 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9\": container with ID starting with 0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9 not found: ID does not exist" containerID="0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.511363 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9"} err="failed to get container status \"0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9\": rpc error: code = NotFound desc = could not find container \"0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9\": container with ID starting with 0316c4e0880e701e3b34f01a7317c0b5be207eb9ee076ad5161a47df180c8de9 not found: ID does not exist" Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.517011 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8fdb65495-jfxkp"] Dec 12 08:13:05 crc kubenswrapper[4867]: I1212 08:13:05.588309 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 12 08:13:05 crc kubenswrapper[4867]: W1212 08:13:05.590569 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72d8033d_f19a_462d_9f0a_1122235ef5e9.slice/crio-d29c4b19ddad27e62789e91bd6b05256c9291659f57ca4e505201cdfb103cc2e WatchSource:0}: Error finding container d29c4b19ddad27e62789e91bd6b05256c9291659f57ca4e505201cdfb103cc2e: Status 404 returned error can't find the container with id d29c4b19ddad27e62789e91bd6b05256c9291659f57ca4e505201cdfb103cc2e Dec 12 08:13:06 crc kubenswrapper[4867]: I1212 08:13:06.470700 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"72d8033d-f19a-462d-9f0a-1122235ef5e9","Type":"ContainerStarted","Data":"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac"} Dec 12 08:13:06 crc kubenswrapper[4867]: I1212 08:13:06.471011 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"72d8033d-f19a-462d-9f0a-1122235ef5e9","Type":"ContainerStarted","Data":"d29c4b19ddad27e62789e91bd6b05256c9291659f57ca4e505201cdfb103cc2e"} Dec 12 08:13:06 crc kubenswrapper[4867]: I1212 08:13:06.493757 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.281629716 podStartE2EDuration="3.493734451s" podCreationTimestamp="2025-12-12 08:13:03 +0000 UTC" firstStartedPulling="2025-12-12 08:13:05.59331188 +0000 UTC m=+5073.164693149" lastFinishedPulling="2025-12-12 08:13:05.805416615 +0000 UTC m=+5073.376797884" observedRunningTime="2025-12-12 08:13:06.48802454 +0000 UTC m=+5074.059405819" watchObservedRunningTime="2025-12-12 08:13:06.493734451 +0000 UTC m=+5074.065115720" Dec 12 08:13:06 crc kubenswrapper[4867]: I1212 08:13:06.849118 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" path="/var/lib/kubelet/pods/f5b343d7-d7cb-4376-be0d-d2f6d229959f/volumes" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.266771 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 12 08:13:11 crc kubenswrapper[4867]: E1212 08:13:11.267603 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="init" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.267622 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="init" Dec 12 08:13:11 crc kubenswrapper[4867]: E1212 08:13:11.267638 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="dnsmasq-dns" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.267646 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="dnsmasq-dns" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.267825 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b343d7-d7cb-4376-be0d-d2f6d229959f" containerName="dnsmasq-dns" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.268951 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.272342 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-k929c" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.272444 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.273363 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.281008 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.288049 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444061 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444167 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctsm5\" (UniqueName: \"kubernetes.io/projected/c21b182d-1e8c-41aa-8fd3-51bc302238de-kube-api-access-ctsm5\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444200 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444251 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444460 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-scripts\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444509 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-config\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.444608 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546214 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctsm5\" (UniqueName: \"kubernetes.io/projected/c21b182d-1e8c-41aa-8fd3-51bc302238de-kube-api-access-ctsm5\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546311 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546346 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546398 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-scripts\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546415 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-config\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546441 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546481 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.546991 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.547558 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-config\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.547652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c21b182d-1e8c-41aa-8fd3-51bc302238de-scripts\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.552068 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.552246 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.553097 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21b182d-1e8c-41aa-8fd3-51bc302238de-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.563714 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctsm5\" (UniqueName: \"kubernetes.io/projected/c21b182d-1e8c-41aa-8fd3-51bc302238de-kube-api-access-ctsm5\") pod \"ovn-northd-0\" (UID: \"c21b182d-1e8c-41aa-8fd3-51bc302238de\") " pod="openstack/ovn-northd-0" Dec 12 08:13:11 crc kubenswrapper[4867]: I1212 08:13:11.595083 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 12 08:13:12 crc kubenswrapper[4867]: I1212 08:13:12.074217 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 12 08:13:12 crc kubenswrapper[4867]: I1212 08:13:12.528121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c21b182d-1e8c-41aa-8fd3-51bc302238de","Type":"ContainerStarted","Data":"bb26d5fc6da0794cbf3977620236580e25f59d74f852b61456ebfe25e4284375"} Dec 12 08:13:12 crc kubenswrapper[4867]: I1212 08:13:12.876421 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:12 crc kubenswrapper[4867]: I1212 08:13:12.930966 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:13:12 crc kubenswrapper[4867]: I1212 08:13:12.931274 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="dnsmasq-dns" containerID="cri-o://e9313c3280f84f2be678dcb7b1936ccd5cee1ecc08176b58bd858e60b6fbbef7" gracePeriod=10 Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.538391 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c21b182d-1e8c-41aa-8fd3-51bc302238de","Type":"ContainerStarted","Data":"063eda483aab2900db0a5007aee10b85930bccd520d093c3a396261dd48a1f70"} Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.538693 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c21b182d-1e8c-41aa-8fd3-51bc302238de","Type":"ContainerStarted","Data":"ffab223f7ad5a34afba694d773c4b041d2c5c359396d9bc508862a1d33a8c9a1"} Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.538708 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.540445 4867 generic.go:334] "Generic (PLEG): container finished" podID="45b18049-f199-4ad0-ad94-64e06149d167" containerID="e9313c3280f84f2be678dcb7b1936ccd5cee1ecc08176b58bd858e60b6fbbef7" exitCode=0 Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.540499 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" event={"ID":"45b18049-f199-4ad0-ad94-64e06149d167","Type":"ContainerDied","Data":"e9313c3280f84f2be678dcb7b1936ccd5cee1ecc08176b58bd858e60b6fbbef7"} Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.555722 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.87884981 podStartE2EDuration="2.555702593s" podCreationTimestamp="2025-12-12 08:13:11 +0000 UTC" firstStartedPulling="2025-12-12 08:13:12.081655311 +0000 UTC m=+5079.653036580" lastFinishedPulling="2025-12-12 08:13:12.758508094 +0000 UTC m=+5080.329889363" observedRunningTime="2025-12-12 08:13:13.554609405 +0000 UTC m=+5081.125990674" watchObservedRunningTime="2025-12-12 08:13:13.555702593 +0000 UTC m=+5081.127083862" Dec 12 08:13:13 crc kubenswrapper[4867]: I1212 08:13:13.909583 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.098921 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc\") pod \"45b18049-f199-4ad0-ad94-64e06149d167\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.099067 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kdbn\" (UniqueName: \"kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn\") pod \"45b18049-f199-4ad0-ad94-64e06149d167\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.099144 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config\") pod \"45b18049-f199-4ad0-ad94-64e06149d167\" (UID: \"45b18049-f199-4ad0-ad94-64e06149d167\") " Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.106402 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn" (OuterVolumeSpecName: "kube-api-access-5kdbn") pod "45b18049-f199-4ad0-ad94-64e06149d167" (UID: "45b18049-f199-4ad0-ad94-64e06149d167"). InnerVolumeSpecName "kube-api-access-5kdbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.136986 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config" (OuterVolumeSpecName: "config") pod "45b18049-f199-4ad0-ad94-64e06149d167" (UID: "45b18049-f199-4ad0-ad94-64e06149d167"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.138263 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45b18049-f199-4ad0-ad94-64e06149d167" (UID: "45b18049-f199-4ad0-ad94-64e06149d167"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.200698 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.201065 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45b18049-f199-4ad0-ad94-64e06149d167-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.201079 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kdbn\" (UniqueName: \"kubernetes.io/projected/45b18049-f199-4ad0-ad94-64e06149d167-kube-api-access-5kdbn\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.548962 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" event={"ID":"45b18049-f199-4ad0-ad94-64e06149d167","Type":"ContainerDied","Data":"ee6e504d4bc9a0e755579848ab8855698ab16a72f701d17aadeffcf337d131aa"} Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.548989 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76958896f7-gv5hr" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.549015 4867 scope.go:117] "RemoveContainer" containerID="e9313c3280f84f2be678dcb7b1936ccd5cee1ecc08176b58bd858e60b6fbbef7" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.570031 4867 scope.go:117] "RemoveContainer" containerID="34c1f2e171b02a46ec310aed4c241d113d2ea43dc63d7da2c5451504ca72972c" Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.581247 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.587932 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76958896f7-gv5hr"] Dec 12 08:13:14 crc kubenswrapper[4867]: I1212 08:13:14.856585 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45b18049-f199-4ad0-ad94-64e06149d167" path="/var/lib/kubelet/pods/45b18049-f199-4ad0-ad94-64e06149d167/volumes" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.346208 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-s95rc"] Dec 12 08:13:16 crc kubenswrapper[4867]: E1212 08:13:16.347940 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="init" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.348028 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="init" Dec 12 08:13:16 crc kubenswrapper[4867]: E1212 08:13:16.348110 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="dnsmasq-dns" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.348176 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="dnsmasq-dns" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.348427 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="45b18049-f199-4ad0-ad94-64e06149d167" containerName="dnsmasq-dns" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.349990 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.360038 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s95rc"] Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.441318 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c222r\" (UniqueName: \"kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.441421 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.448752 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-691b-account-create-update-8589s"] Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.450122 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.452187 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.464399 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-691b-account-create-update-8589s"] Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.542835 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.542899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwfn6\" (UniqueName: \"kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.542970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c222r\" (UniqueName: \"kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.543031 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.543797 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.563586 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c222r\" (UniqueName: \"kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r\") pod \"keystone-db-create-s95rc\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.644816 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.644884 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwfn6\" (UniqueName: \"kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.645980 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.662039 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwfn6\" (UniqueName: \"kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6\") pod \"keystone-691b-account-create-update-8589s\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.676180 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:16 crc kubenswrapper[4867]: I1212 08:13:16.817275 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.113503 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s95rc"] Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.253067 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-691b-account-create-update-8589s"] Dec 12 08:13:17 crc kubenswrapper[4867]: W1212 08:13:17.275388 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce280bcc_4651_4777_8e50_f6951e781224.slice/crio-00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958 WatchSource:0}: Error finding container 00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958: Status 404 returned error can't find the container with id 00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958 Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.577601 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-691b-account-create-update-8589s" event={"ID":"ce280bcc-4651-4777-8e50-f6951e781224","Type":"ContainerStarted","Data":"3e7bd986c549be106a8d6ee6bd71df5784a614dba90580d4ad26b0795ab4e217"} Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.577726 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-691b-account-create-update-8589s" event={"ID":"ce280bcc-4651-4777-8e50-f6951e781224","Type":"ContainerStarted","Data":"00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958"} Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.579702 4867 generic.go:334] "Generic (PLEG): container finished" podID="fbfa62db-247f-44d3-bd2b-efb127cb954e" containerID="83f0eabf639e422848a12e485655f4d6f22a4fb3304c77654d9ca0c6d4d445d8" exitCode=0 Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.579762 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s95rc" event={"ID":"fbfa62db-247f-44d3-bd2b-efb127cb954e","Type":"ContainerDied","Data":"83f0eabf639e422848a12e485655f4d6f22a4fb3304c77654d9ca0c6d4d445d8"} Dec 12 08:13:17 crc kubenswrapper[4867]: I1212 08:13:17.579827 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s95rc" event={"ID":"fbfa62db-247f-44d3-bd2b-efb127cb954e","Type":"ContainerStarted","Data":"7e5a87e4ea25cc90f41676945f3420a042d237e1e66861de4d03b3ef2f2310bb"} Dec 12 08:13:18 crc kubenswrapper[4867]: I1212 08:13:18.587949 4867 generic.go:334] "Generic (PLEG): container finished" podID="ce280bcc-4651-4777-8e50-f6951e781224" containerID="3e7bd986c549be106a8d6ee6bd71df5784a614dba90580d4ad26b0795ab4e217" exitCode=0 Dec 12 08:13:18 crc kubenswrapper[4867]: I1212 08:13:18.588040 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-691b-account-create-update-8589s" event={"ID":"ce280bcc-4651-4777-8e50-f6951e781224","Type":"ContainerDied","Data":"3e7bd986c549be106a8d6ee6bd71df5784a614dba90580d4ad26b0795ab4e217"} Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.002199 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.008051 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.184832 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c222r\" (UniqueName: \"kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r\") pod \"fbfa62db-247f-44d3-bd2b-efb127cb954e\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.185035 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwfn6\" (UniqueName: \"kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6\") pod \"ce280bcc-4651-4777-8e50-f6951e781224\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.185089 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts\") pod \"ce280bcc-4651-4777-8e50-f6951e781224\" (UID: \"ce280bcc-4651-4777-8e50-f6951e781224\") " Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.185151 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts\") pod \"fbfa62db-247f-44d3-bd2b-efb127cb954e\" (UID: \"fbfa62db-247f-44d3-bd2b-efb127cb954e\") " Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.186043 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fbfa62db-247f-44d3-bd2b-efb127cb954e" (UID: "fbfa62db-247f-44d3-bd2b-efb127cb954e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.187389 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce280bcc-4651-4777-8e50-f6951e781224" (UID: "ce280bcc-4651-4777-8e50-f6951e781224"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.192495 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r" (OuterVolumeSpecName: "kube-api-access-c222r") pod "fbfa62db-247f-44d3-bd2b-efb127cb954e" (UID: "fbfa62db-247f-44d3-bd2b-efb127cb954e"). InnerVolumeSpecName "kube-api-access-c222r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.192607 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6" (OuterVolumeSpecName: "kube-api-access-lwfn6") pod "ce280bcc-4651-4777-8e50-f6951e781224" (UID: "ce280bcc-4651-4777-8e50-f6951e781224"). InnerVolumeSpecName "kube-api-access-lwfn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.286899 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c222r\" (UniqueName: \"kubernetes.io/projected/fbfa62db-247f-44d3-bd2b-efb127cb954e-kube-api-access-c222r\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.286930 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwfn6\" (UniqueName: \"kubernetes.io/projected/ce280bcc-4651-4777-8e50-f6951e781224-kube-api-access-lwfn6\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.286941 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce280bcc-4651-4777-8e50-f6951e781224-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.286950 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fbfa62db-247f-44d3-bd2b-efb127cb954e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.596206 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s95rc" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.596183 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s95rc" event={"ID":"fbfa62db-247f-44d3-bd2b-efb127cb954e","Type":"ContainerDied","Data":"7e5a87e4ea25cc90f41676945f3420a042d237e1e66861de4d03b3ef2f2310bb"} Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.596536 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e5a87e4ea25cc90f41676945f3420a042d237e1e66861de4d03b3ef2f2310bb" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.597641 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-691b-account-create-update-8589s" event={"ID":"ce280bcc-4651-4777-8e50-f6951e781224","Type":"ContainerDied","Data":"00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958"} Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.597669 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00bed3747ce7c24a67f6bc0f5c0ce74139b3a20f412649fb0e1b1672d4e71958" Dec 12 08:13:19 crc kubenswrapper[4867]: I1212 08:13:19.597705 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-691b-account-create-update-8589s" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.013378 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-tgxgc"] Dec 12 08:13:22 crc kubenswrapper[4867]: E1212 08:13:22.014138 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbfa62db-247f-44d3-bd2b-efb127cb954e" containerName="mariadb-database-create" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.014154 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbfa62db-247f-44d3-bd2b-efb127cb954e" containerName="mariadb-database-create" Dec 12 08:13:22 crc kubenswrapper[4867]: E1212 08:13:22.014187 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce280bcc-4651-4777-8e50-f6951e781224" containerName="mariadb-account-create-update" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.014195 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce280bcc-4651-4777-8e50-f6951e781224" containerName="mariadb-account-create-update" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.014427 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce280bcc-4651-4777-8e50-f6951e781224" containerName="mariadb-account-create-update" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.014447 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbfa62db-247f-44d3-bd2b-efb127cb954e" containerName="mariadb-database-create" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.015141 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.017608 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.017613 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.017831 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6blzp" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.018058 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.026432 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tgxgc"] Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.127915 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.128155 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl6qq\" (UniqueName: \"kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.128189 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.230004 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.230118 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl6qq\" (UniqueName: \"kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.230144 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.234868 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.234884 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.248136 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl6qq\" (UniqueName: \"kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq\") pod \"keystone-db-sync-tgxgc\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.333495 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:22 crc kubenswrapper[4867]: I1212 08:13:22.766211 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tgxgc"] Dec 12 08:13:23 crc kubenswrapper[4867]: I1212 08:13:23.631653 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tgxgc" event={"ID":"f8915409-bbc7-4db4-9a74-0bbfe956f80f","Type":"ContainerStarted","Data":"287ff11caef49a51777118ac00ab2141270e73f8c51c49fa13e3f7a6dfbc0e36"} Dec 12 08:13:26 crc kubenswrapper[4867]: I1212 08:13:26.673572 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 12 08:13:29 crc kubenswrapper[4867]: I1212 08:13:29.698459 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tgxgc" event={"ID":"f8915409-bbc7-4db4-9a74-0bbfe956f80f","Type":"ContainerStarted","Data":"2b6e73bd3ad851f8f4e007959a4e7d88246b829788dde949727ae804af2d1de7"} Dec 12 08:13:29 crc kubenswrapper[4867]: I1212 08:13:29.711857 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-tgxgc" podStartSLOduration=2.865392366 podStartE2EDuration="8.711840635s" podCreationTimestamp="2025-12-12 08:13:21 +0000 UTC" firstStartedPulling="2025-12-12 08:13:22.773878158 +0000 UTC m=+5090.345259427" lastFinishedPulling="2025-12-12 08:13:28.620326427 +0000 UTC m=+5096.191707696" observedRunningTime="2025-12-12 08:13:29.711112387 +0000 UTC m=+5097.282493666" watchObservedRunningTime="2025-12-12 08:13:29.711840635 +0000 UTC m=+5097.283221904" Dec 12 08:13:30 crc kubenswrapper[4867]: I1212 08:13:30.706872 4867 generic.go:334] "Generic (PLEG): container finished" podID="f8915409-bbc7-4db4-9a74-0bbfe956f80f" containerID="2b6e73bd3ad851f8f4e007959a4e7d88246b829788dde949727ae804af2d1de7" exitCode=0 Dec 12 08:13:30 crc kubenswrapper[4867]: I1212 08:13:30.706913 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tgxgc" event={"ID":"f8915409-bbc7-4db4-9a74-0bbfe956f80f","Type":"ContainerDied","Data":"2b6e73bd3ad851f8f4e007959a4e7d88246b829788dde949727ae804af2d1de7"} Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.070901 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.213025 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl6qq\" (UniqueName: \"kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq\") pod \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.213264 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle\") pod \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.213293 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data\") pod \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\" (UID: \"f8915409-bbc7-4db4-9a74-0bbfe956f80f\") " Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.221033 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq" (OuterVolumeSpecName: "kube-api-access-gl6qq") pod "f8915409-bbc7-4db4-9a74-0bbfe956f80f" (UID: "f8915409-bbc7-4db4-9a74-0bbfe956f80f"). InnerVolumeSpecName "kube-api-access-gl6qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.240655 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8915409-bbc7-4db4-9a74-0bbfe956f80f" (UID: "f8915409-bbc7-4db4-9a74-0bbfe956f80f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.256348 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data" (OuterVolumeSpecName: "config-data") pod "f8915409-bbc7-4db4-9a74-0bbfe956f80f" (UID: "f8915409-bbc7-4db4-9a74-0bbfe956f80f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.315768 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.315806 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl6qq\" (UniqueName: \"kubernetes.io/projected/f8915409-bbc7-4db4-9a74-0bbfe956f80f-kube-api-access-gl6qq\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.315816 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8915409-bbc7-4db4-9a74-0bbfe956f80f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.723650 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tgxgc" event={"ID":"f8915409-bbc7-4db4-9a74-0bbfe956f80f","Type":"ContainerDied","Data":"287ff11caef49a51777118ac00ab2141270e73f8c51c49fa13e3f7a6dfbc0e36"} Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.723690 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="287ff11caef49a51777118ac00ab2141270e73f8c51c49fa13e3f7a6dfbc0e36" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.723957 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tgxgc" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.963920 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:13:32 crc kubenswrapper[4867]: E1212 08:13:32.964571 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8915409-bbc7-4db4-9a74-0bbfe956f80f" containerName="keystone-db-sync" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.964590 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8915409-bbc7-4db4-9a74-0bbfe956f80f" containerName="keystone-db-sync" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.964763 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8915409-bbc7-4db4-9a74-0bbfe956f80f" containerName="keystone-db-sync" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.965969 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:32 crc kubenswrapper[4867]: I1212 08:13:32.974543 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.028433 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.028528 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.028756 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzljc\" (UniqueName: \"kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.028836 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.028890 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.037918 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2r8v4"] Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.039471 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.044793 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.044835 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.044876 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.044835 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6blzp" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.045848 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.052260 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2r8v4"] Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130096 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130143 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqhc7\" (UniqueName: \"kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130198 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzljc\" (UniqueName: \"kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130285 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130321 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130349 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130368 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130432 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130475 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130490 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.130522 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.134040 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.134706 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.134710 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.135698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.148274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzljc\" (UniqueName: \"kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc\") pod \"dnsmasq-dns-76d5f75dd7-ptnt2\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231602 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231671 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231707 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231754 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231795 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.231813 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqhc7\" (UniqueName: \"kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.234914 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.235057 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.236461 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.236518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.237073 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.254964 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqhc7\" (UniqueName: \"kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7\") pod \"keystone-bootstrap-2r8v4\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.292154 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.356700 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.613114 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:13:33 crc kubenswrapper[4867]: W1212 08:13:33.630573 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod036afb2d_c0e8_4f31_b0b8_e5548773af3c.slice/crio-e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591 WatchSource:0}: Error finding container e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591: Status 404 returned error can't find the container with id e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591 Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.742324 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" event={"ID":"036afb2d-c0e8-4f31-b0b8-e5548773af3c","Type":"ContainerStarted","Data":"e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591"} Dec 12 08:13:33 crc kubenswrapper[4867]: I1212 08:13:33.978401 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2r8v4"] Dec 12 08:13:34 crc kubenswrapper[4867]: I1212 08:13:34.751270 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r8v4" event={"ID":"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228","Type":"ContainerStarted","Data":"a48842721da335fda06794b72126d6dac1510a9b06123f879baad397e6cde37d"} Dec 12 08:13:34 crc kubenswrapper[4867]: I1212 08:13:34.751609 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r8v4" event={"ID":"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228","Type":"ContainerStarted","Data":"4469b16ec52a58586ebf99ce11d41bca507609aab123d104fdab6ee2ab58c2d1"} Dec 12 08:13:34 crc kubenswrapper[4867]: I1212 08:13:34.752638 4867 generic.go:334] "Generic (PLEG): container finished" podID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerID="008070a9255f964ecb06601f113844333f4d790b3bc7119122f33b9648090e9c" exitCode=0 Dec 12 08:13:34 crc kubenswrapper[4867]: I1212 08:13:34.752683 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" event={"ID":"036afb2d-c0e8-4f31-b0b8-e5548773af3c","Type":"ContainerDied","Data":"008070a9255f964ecb06601f113844333f4d790b3bc7119122f33b9648090e9c"} Dec 12 08:13:34 crc kubenswrapper[4867]: I1212 08:13:34.804418 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2r8v4" podStartSLOduration=1.8043969340000001 podStartE2EDuration="1.804396934s" podCreationTimestamp="2025-12-12 08:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:34.776137648 +0000 UTC m=+5102.347518917" watchObservedRunningTime="2025-12-12 08:13:34.804396934 +0000 UTC m=+5102.375778213" Dec 12 08:13:35 crc kubenswrapper[4867]: I1212 08:13:35.775498 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" event={"ID":"036afb2d-c0e8-4f31-b0b8-e5548773af3c","Type":"ContainerStarted","Data":"eb9a1029e25d39c84f7a191d1a9ced6a9c9703852deb7449b67da9801e0a227a"} Dec 12 08:13:35 crc kubenswrapper[4867]: I1212 08:13:35.775798 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:35 crc kubenswrapper[4867]: I1212 08:13:35.805296 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" podStartSLOduration=3.805275179 podStartE2EDuration="3.805275179s" podCreationTimestamp="2025-12-12 08:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:35.799889686 +0000 UTC m=+5103.371270975" watchObservedRunningTime="2025-12-12 08:13:35.805275179 +0000 UTC m=+5103.376656448" Dec 12 08:13:37 crc kubenswrapper[4867]: I1212 08:13:37.789185 4867 generic.go:334] "Generic (PLEG): container finished" podID="f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" containerID="a48842721da335fda06794b72126d6dac1510a9b06123f879baad397e6cde37d" exitCode=0 Dec 12 08:13:37 crc kubenswrapper[4867]: I1212 08:13:37.789296 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r8v4" event={"ID":"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228","Type":"ContainerDied","Data":"a48842721da335fda06794b72126d6dac1510a9b06123f879baad397e6cde37d"} Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.182525 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226202 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226591 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226672 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226725 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqhc7\" (UniqueName: \"kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226769 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.226834 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data\") pod \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\" (UID: \"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228\") " Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.231864 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts" (OuterVolumeSpecName: "scripts") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.232311 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7" (OuterVolumeSpecName: "kube-api-access-xqhc7") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "kube-api-access-xqhc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.232392 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.233117 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.250277 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data" (OuterVolumeSpecName: "config-data") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.253028 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" (UID: "f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328827 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328868 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328881 4867 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328896 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqhc7\" (UniqueName: \"kubernetes.io/projected/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-kube-api-access-xqhc7\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328909 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.328919 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.809957 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r8v4" event={"ID":"f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228","Type":"ContainerDied","Data":"4469b16ec52a58586ebf99ce11d41bca507609aab123d104fdab6ee2ab58c2d1"} Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.810002 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4469b16ec52a58586ebf99ce11d41bca507609aab123d104fdab6ee2ab58c2d1" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.810069 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r8v4" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.890793 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2r8v4"] Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.898839 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2r8v4"] Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.975712 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bfpsl"] Dec 12 08:13:39 crc kubenswrapper[4867]: E1212 08:13:39.976279 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" containerName="keystone-bootstrap" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.976353 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" containerName="keystone-bootstrap" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.976573 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" containerName="keystone-bootstrap" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.977180 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.984685 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bfpsl"] Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.999485 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.999590 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.999636 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6blzp" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.999726 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 08:13:39 crc kubenswrapper[4867]: I1212 08:13:39.999956 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.047284 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vh5h\" (UniqueName: \"kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.047777 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.047939 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.048034 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.048111 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.048257 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.149327 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.149656 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.149764 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.149847 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.149916 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.150015 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vh5h\" (UniqueName: \"kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.155869 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.156787 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.157398 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.157936 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.168306 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.171887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vh5h\" (UniqueName: \"kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h\") pod \"keystone-bootstrap-bfpsl\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.312398 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.758640 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bfpsl"] Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.819070 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfpsl" event={"ID":"0fc25c26-edca-4bc5-861a-f66e643cf7f0","Type":"ContainerStarted","Data":"44b2e735f1754001fa99a6b4cd9902bf88eb899555b2e09bed617d10a85eeaae"} Dec 12 08:13:40 crc kubenswrapper[4867]: I1212 08:13:40.849689 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228" path="/var/lib/kubelet/pods/f53d0aaa-cfd5-4c66-bb9f-ccfb26dbc228/volumes" Dec 12 08:13:41 crc kubenswrapper[4867]: I1212 08:13:41.829149 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfpsl" event={"ID":"0fc25c26-edca-4bc5-861a-f66e643cf7f0","Type":"ContainerStarted","Data":"dafc4a0d2c9fede68c1bd03410af69ccdd4ebe7141380fcf13a84f35f43a2244"} Dec 12 08:13:41 crc kubenswrapper[4867]: I1212 08:13:41.852040 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bfpsl" podStartSLOduration=2.852019051 podStartE2EDuration="2.852019051s" podCreationTimestamp="2025-12-12 08:13:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:41.845898471 +0000 UTC m=+5109.417279750" watchObservedRunningTime="2025-12-12 08:13:41.852019051 +0000 UTC m=+5109.423400320" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.294477 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.351187 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.351442 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="dnsmasq-dns" containerID="cri-o://7525e1e93a72e8fb116d175998e511e0b9a6955a80e82c3df82426e39d1e7c4e" gracePeriod=10 Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.852780 4867 generic.go:334] "Generic (PLEG): container finished" podID="0fc25c26-edca-4bc5-861a-f66e643cf7f0" containerID="dafc4a0d2c9fede68c1bd03410af69ccdd4ebe7141380fcf13a84f35f43a2244" exitCode=0 Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.852905 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfpsl" event={"ID":"0fc25c26-edca-4bc5-861a-f66e643cf7f0","Type":"ContainerDied","Data":"dafc4a0d2c9fede68c1bd03410af69ccdd4ebe7141380fcf13a84f35f43a2244"} Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.863811 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerID="7525e1e93a72e8fb116d175998e511e0b9a6955a80e82c3df82426e39d1e7c4e" exitCode=0 Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.863886 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" event={"ID":"d1e6b2fd-8b09-4498-9fbe-c1b27625f264","Type":"ContainerDied","Data":"7525e1e93a72e8fb116d175998e511e0b9a6955a80e82c3df82426e39d1e7c4e"} Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.864327 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" event={"ID":"d1e6b2fd-8b09-4498-9fbe-c1b27625f264","Type":"ContainerDied","Data":"9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c"} Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.864343 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9012d54e24d19f089708b31c50a9ea3fb271c19bc2c57cf293371b04014b664c" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.876068 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.923342 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc\") pod \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.923625 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb\") pod \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.923924 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config\") pod \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.924106 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb\") pod \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.924303 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzz6q\" (UniqueName: \"kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q\") pod \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\" (UID: \"d1e6b2fd-8b09-4498-9fbe-c1b27625f264\") " Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.929706 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q" (OuterVolumeSpecName: "kube-api-access-zzz6q") pod "d1e6b2fd-8b09-4498-9fbe-c1b27625f264" (UID: "d1e6b2fd-8b09-4498-9fbe-c1b27625f264"). InnerVolumeSpecName "kube-api-access-zzz6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.975926 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d1e6b2fd-8b09-4498-9fbe-c1b27625f264" (UID: "d1e6b2fd-8b09-4498-9fbe-c1b27625f264"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.978210 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config" (OuterVolumeSpecName: "config") pod "d1e6b2fd-8b09-4498-9fbe-c1b27625f264" (UID: "d1e6b2fd-8b09-4498-9fbe-c1b27625f264"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:43 crc kubenswrapper[4867]: I1212 08:13:43.983512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d1e6b2fd-8b09-4498-9fbe-c1b27625f264" (UID: "d1e6b2fd-8b09-4498-9fbe-c1b27625f264"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.000801 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1e6b2fd-8b09-4498-9fbe-c1b27625f264" (UID: "d1e6b2fd-8b09-4498-9fbe-c1b27625f264"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.027566 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzz6q\" (UniqueName: \"kubernetes.io/projected/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-kube-api-access-zzz6q\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.027632 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.027650 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.027663 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.027674 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1e6b2fd-8b09-4498-9fbe-c1b27625f264-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.870706 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd56d8775-cfgzb" Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.897083 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:44 crc kubenswrapper[4867]: I1212 08:13:44.902881 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cd56d8775-cfgzb"] Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.241135 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354005 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354088 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354113 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354172 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vh5h\" (UniqueName: \"kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354222 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.354349 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys\") pod \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\" (UID: \"0fc25c26-edca-4bc5-861a-f66e643cf7f0\") " Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.359685 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.364588 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.366283 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h" (OuterVolumeSpecName: "kube-api-access-7vh5h") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "kube-api-access-7vh5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.366396 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts" (OuterVolumeSpecName: "scripts") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.380437 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data" (OuterVolumeSpecName: "config-data") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.381380 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fc25c26-edca-4bc5-861a-f66e643cf7f0" (UID: "0fc25c26-edca-4bc5-861a-f66e643cf7f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456351 4867 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456388 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456396 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456405 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456414 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fc25c26-edca-4bc5-861a-f66e643cf7f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.456424 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vh5h\" (UniqueName: \"kubernetes.io/projected/0fc25c26-edca-4bc5-861a-f66e643cf7f0-kube-api-access-7vh5h\") on node \"crc\" DevicePath \"\"" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.880340 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bfpsl" event={"ID":"0fc25c26-edca-4bc5-861a-f66e643cf7f0","Type":"ContainerDied","Data":"44b2e735f1754001fa99a6b4cd9902bf88eb899555b2e09bed617d10a85eeaae"} Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.880381 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44b2e735f1754001fa99a6b4cd9902bf88eb899555b2e09bed617d10a85eeaae" Dec 12 08:13:45 crc kubenswrapper[4867]: I1212 08:13:45.880432 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bfpsl" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.062998 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5789865856-hkfwx"] Dec 12 08:13:46 crc kubenswrapper[4867]: E1212 08:13:46.063383 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fc25c26-edca-4bc5-861a-f66e643cf7f0" containerName="keystone-bootstrap" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.063400 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fc25c26-edca-4bc5-861a-f66e643cf7f0" containerName="keystone-bootstrap" Dec 12 08:13:46 crc kubenswrapper[4867]: E1212 08:13:46.063410 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="dnsmasq-dns" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.063417 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="dnsmasq-dns" Dec 12 08:13:46 crc kubenswrapper[4867]: E1212 08:13:46.063448 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="init" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.063454 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="init" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.063617 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fc25c26-edca-4bc5-861a-f66e643cf7f0" containerName="keystone-bootstrap" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.063642 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" containerName="dnsmasq-dns" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.064176 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.067286 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.067404 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.067521 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.067564 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.068164 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-6blzp" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.068394 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.092137 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5789865856-hkfwx"] Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170707 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-internal-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-config-data\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170831 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdlp2\" (UniqueName: \"kubernetes.io/projected/864c3801-1b19-46fb-8564-8e2d772963bc-kube-api-access-rdlp2\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170857 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-credential-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170884 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-scripts\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170920 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-combined-ca-bundle\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170945 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-fernet-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.170983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-public-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272278 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-combined-ca-bundle\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272342 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-fernet-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272402 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-public-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272445 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-internal-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272482 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-config-data\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272526 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdlp2\" (UniqueName: \"kubernetes.io/projected/864c3801-1b19-46fb-8564-8e2d772963bc-kube-api-access-rdlp2\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-credential-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.272703 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-scripts\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.278185 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-fernet-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.279147 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-public-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.283657 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-combined-ca-bundle\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.284098 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-scripts\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.284418 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-config-data\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.286819 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-internal-tls-certs\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.293520 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdlp2\" (UniqueName: \"kubernetes.io/projected/864c3801-1b19-46fb-8564-8e2d772963bc-kube-api-access-rdlp2\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.298833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/864c3801-1b19-46fb-8564-8e2d772963bc-credential-keys\") pod \"keystone-5789865856-hkfwx\" (UID: \"864c3801-1b19-46fb-8564-8e2d772963bc\") " pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.388866 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.806856 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5789865856-hkfwx"] Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.848264 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1e6b2fd-8b09-4498-9fbe-c1b27625f264" path="/var/lib/kubelet/pods/d1e6b2fd-8b09-4498-9fbe-c1b27625f264/volumes" Dec 12 08:13:46 crc kubenswrapper[4867]: I1212 08:13:46.898120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5789865856-hkfwx" event={"ID":"864c3801-1b19-46fb-8564-8e2d772963bc","Type":"ContainerStarted","Data":"123063053e0c9b41e7f63a8c8a504653d139f87f8debaee3c1dcbca3f296ed99"} Dec 12 08:13:47 crc kubenswrapper[4867]: I1212 08:13:47.906777 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5789865856-hkfwx" event={"ID":"864c3801-1b19-46fb-8564-8e2d772963bc","Type":"ContainerStarted","Data":"b3316a36369e830f5053eec88e9f159609f33ceef46acbf8cd417ee1a4e5413d"} Dec 12 08:13:47 crc kubenswrapper[4867]: I1212 08:13:47.907135 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:13:47 crc kubenswrapper[4867]: I1212 08:13:47.927429 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5789865856-hkfwx" podStartSLOduration=1.927407439 podStartE2EDuration="1.927407439s" podCreationTimestamp="2025-12-12 08:13:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:13:47.921679248 +0000 UTC m=+5115.493060537" watchObservedRunningTime="2025-12-12 08:13:47.927407439 +0000 UTC m=+5115.498788708" Dec 12 08:14:17 crc kubenswrapper[4867]: I1212 08:14:17.958929 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5789865856-hkfwx" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.368381 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.370074 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.374913 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.375035 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.375176 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-66zxh" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.381933 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.425783 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: E1212 08:14:21.430839 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-45ltw openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="fcccd85b-855e-4c17-9430-85b1fc566e93" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.446450 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.456604 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.458743 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.469424 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.469505 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.469632 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.469677 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45ltw\" (UniqueName: \"kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.478053 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.570915 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571077 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l252f\" (UniqueName: \"kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571142 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571180 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571197 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571353 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ltw\" (UniqueName: \"kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.571434 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.572487 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: E1212 08:14:21.574144 4867 projected.go:194] Error preparing data for projected volume kube-api-access-45ltw for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fcccd85b-855e-4c17-9430-85b1fc566e93) does not match the UID in record. The object might have been deleted and then recreated Dec 12 08:14:21 crc kubenswrapper[4867]: E1212 08:14:21.574210 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw podName:fcccd85b-855e-4c17-9430-85b1fc566e93 nodeName:}" failed. No retries permitted until 2025-12-12 08:14:22.074190346 +0000 UTC m=+5149.645571615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-45ltw" (UniqueName: "kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw") pod "openstackclient" (UID: "fcccd85b-855e-4c17-9430-85b1fc566e93") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fcccd85b-855e-4c17-9430-85b1fc566e93) does not match the UID in record. The object might have been deleted and then recreated Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.578891 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.579887 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.673137 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l252f\" (UniqueName: \"kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.673217 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.673382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.673479 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.674523 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.681137 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.681212 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.691059 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l252f\" (UniqueName: \"kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f\") pod \"openstackclient\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " pod="openstack/openstackclient" Dec 12 08:14:21 crc kubenswrapper[4867]: I1212 08:14:21.780256 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.080559 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45ltw\" (UniqueName: \"kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw\") pod \"openstackclient\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " pod="openstack/openstackclient" Dec 12 08:14:22 crc kubenswrapper[4867]: E1212 08:14:22.082846 4867 projected.go:194] Error preparing data for projected volume kube-api-access-45ltw for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fcccd85b-855e-4c17-9430-85b1fc566e93) does not match the UID in record. The object might have been deleted and then recreated Dec 12 08:14:22 crc kubenswrapper[4867]: E1212 08:14:22.082914 4867 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw podName:fcccd85b-855e-4c17-9430-85b1fc566e93 nodeName:}" failed. No retries permitted until 2025-12-12 08:14:23.082899158 +0000 UTC m=+5150.654280427 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-45ltw" (UniqueName: "kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw") pod "openstackclient" (UID: "fcccd85b-855e-4c17-9430-85b1fc566e93") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (fcccd85b-855e-4c17-9430-85b1fc566e93) does not match the UID in record. The object might have been deleted and then recreated Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.177750 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.182930 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fcccd85b-855e-4c17-9430-85b1fc566e93" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.189495 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.247984 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.254324 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.289315 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle\") pod \"fcccd85b-855e-4c17-9430-85b1fc566e93\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.290345 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret\") pod \"fcccd85b-855e-4c17-9430-85b1fc566e93\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.290406 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config\") pod \"fcccd85b-855e-4c17-9430-85b1fc566e93\" (UID: \"fcccd85b-855e-4c17-9430-85b1fc566e93\") " Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.291130 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fcccd85b-855e-4c17-9430-85b1fc566e93" (UID: "fcccd85b-855e-4c17-9430-85b1fc566e93"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.291669 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.291689 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45ltw\" (UniqueName: \"kubernetes.io/projected/fcccd85b-855e-4c17-9430-85b1fc566e93-kube-api-access-45ltw\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.295361 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcccd85b-855e-4c17-9430-85b1fc566e93" (UID: "fcccd85b-855e-4c17-9430-85b1fc566e93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.295679 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fcccd85b-855e-4c17-9430-85b1fc566e93" (UID: "fcccd85b-855e-4c17-9430-85b1fc566e93"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.392800 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.392839 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fcccd85b-855e-4c17-9430-85b1fc566e93-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:22 crc kubenswrapper[4867]: I1212 08:14:22.856882 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcccd85b-855e-4c17-9430-85b1fc566e93" path="/var/lib/kubelet/pods/fcccd85b-855e-4c17-9430-85b1fc566e93/volumes" Dec 12 08:14:23 crc kubenswrapper[4867]: I1212 08:14:23.186251 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"07ec8a92-18c0-4dae-9f03-775fe99bcb01","Type":"ContainerStarted","Data":"4e91a66f00a2901b169ca10457651fba361fb1c1ecd455643e995bd403a26a59"} Dec 12 08:14:23 crc kubenswrapper[4867]: I1212 08:14:23.186282 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:14:23 crc kubenswrapper[4867]: I1212 08:14:23.194463 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fcccd85b-855e-4c17-9430-85b1fc566e93" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.426015 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.428405 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.433629 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.549107 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.549191 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.549313 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws97g\" (UniqueName: \"kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.650790 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws97g\" (UniqueName: \"kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.650895 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.650927 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.651398 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.651889 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.673341 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws97g\" (UniqueName: \"kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g\") pod \"certified-operators-7njx8\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:31 crc kubenswrapper[4867]: I1212 08:14:31.757001 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.044642 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.280935 4867 generic.go:334] "Generic (PLEG): container finished" podID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerID="cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9" exitCode=0 Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.280995 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerDied","Data":"cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9"} Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.281055 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerStarted","Data":"3123b7fbd70ea8782e86bd9b04f8b12d4a452e07b612f1d65c09fabf5b588a49"} Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.283110 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"07ec8a92-18c0-4dae-9f03-775fe99bcb01","Type":"ContainerStarted","Data":"0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd"} Dec 12 08:14:34 crc kubenswrapper[4867]: I1212 08:14:34.319355 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.898679883 podStartE2EDuration="13.319334855s" podCreationTimestamp="2025-12-12 08:14:21 +0000 UTC" firstStartedPulling="2025-12-12 08:14:22.254098535 +0000 UTC m=+5149.825479804" lastFinishedPulling="2025-12-12 08:14:33.674753487 +0000 UTC m=+5161.246134776" observedRunningTime="2025-12-12 08:14:34.316914735 +0000 UTC m=+5161.888296004" watchObservedRunningTime="2025-12-12 08:14:34.319334855 +0000 UTC m=+5161.890716124" Dec 12 08:14:36 crc kubenswrapper[4867]: I1212 08:14:36.302512 4867 generic.go:334] "Generic (PLEG): container finished" podID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerID="2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60" exitCode=0 Dec 12 08:14:36 crc kubenswrapper[4867]: I1212 08:14:36.302657 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerDied","Data":"2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60"} Dec 12 08:14:37 crc kubenswrapper[4867]: I1212 08:14:37.313349 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerStarted","Data":"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32"} Dec 12 08:14:37 crc kubenswrapper[4867]: I1212 08:14:37.336889 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7njx8" podStartSLOduration=3.570600184 podStartE2EDuration="6.336871957s" podCreationTimestamp="2025-12-12 08:14:31 +0000 UTC" firstStartedPulling="2025-12-12 08:14:34.282520479 +0000 UTC m=+5161.853901748" lastFinishedPulling="2025-12-12 08:14:37.048792252 +0000 UTC m=+5164.620173521" observedRunningTime="2025-12-12 08:14:37.333135116 +0000 UTC m=+5164.904516385" watchObservedRunningTime="2025-12-12 08:14:37.336871957 +0000 UTC m=+5164.908253226" Dec 12 08:14:41 crc kubenswrapper[4867]: I1212 08:14:41.758037 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:41 crc kubenswrapper[4867]: I1212 08:14:41.758502 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:41 crc kubenswrapper[4867]: I1212 08:14:41.801319 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:42 crc kubenswrapper[4867]: I1212 08:14:42.421747 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:42 crc kubenswrapper[4867]: I1212 08:14:42.492948 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:44 crc kubenswrapper[4867]: I1212 08:14:44.371600 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7njx8" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="registry-server" containerID="cri-o://b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32" gracePeriod=2 Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.310792 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.384900 4867 generic.go:334] "Generic (PLEG): container finished" podID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerID="b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32" exitCode=0 Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.385134 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerDied","Data":"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32"} Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.386992 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7njx8" event={"ID":"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d","Type":"ContainerDied","Data":"3123b7fbd70ea8782e86bd9b04f8b12d4a452e07b612f1d65c09fabf5b588a49"} Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.387088 4867 scope.go:117] "RemoveContainer" containerID="b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.385272 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7njx8" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.414375 4867 scope.go:117] "RemoveContainer" containerID="2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.435916 4867 scope.go:117] "RemoveContainer" containerID="cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.482672 4867 scope.go:117] "RemoveContainer" containerID="b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32" Dec 12 08:14:45 crc kubenswrapper[4867]: E1212 08:14:45.484830 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32\": container with ID starting with b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32 not found: ID does not exist" containerID="b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.484889 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32"} err="failed to get container status \"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32\": rpc error: code = NotFound desc = could not find container \"b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32\": container with ID starting with b5dc58be4c2225c24e4322a4a46f5b02850d8add2c6979a8ab691502ca4a9d32 not found: ID does not exist" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.484926 4867 scope.go:117] "RemoveContainer" containerID="2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60" Dec 12 08:14:45 crc kubenswrapper[4867]: E1212 08:14:45.485277 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60\": container with ID starting with 2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60 not found: ID does not exist" containerID="2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.485335 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60"} err="failed to get container status \"2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60\": rpc error: code = NotFound desc = could not find container \"2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60\": container with ID starting with 2c45419057761cd6c4d1525c3fef98be349d4a20b41ab8406b82badf34048a60 not found: ID does not exist" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.485374 4867 scope.go:117] "RemoveContainer" containerID="cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9" Dec 12 08:14:45 crc kubenswrapper[4867]: E1212 08:14:45.485896 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9\": container with ID starting with cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9 not found: ID does not exist" containerID="cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.485940 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9"} err="failed to get container status \"cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9\": rpc error: code = NotFound desc = could not find container \"cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9\": container with ID starting with cf6f345c4024929302503aaad03cbb7ad21222033d0467219c1d386c3b39d1b9 not found: ID does not exist" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.498772 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws97g\" (UniqueName: \"kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g\") pod \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.498875 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities\") pod \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.498954 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content\") pod \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\" (UID: \"a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d\") " Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.502831 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities" (OuterVolumeSpecName: "utilities") pod "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" (UID: "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.508387 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g" (OuterVolumeSpecName: "kube-api-access-ws97g") pod "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" (UID: "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d"). InnerVolumeSpecName "kube-api-access-ws97g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.556184 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" (UID: "a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.602912 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws97g\" (UniqueName: \"kubernetes.io/projected/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-kube-api-access-ws97g\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.602944 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.602954 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.724307 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:45 crc kubenswrapper[4867]: I1212 08:14:45.733695 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7njx8"] Dec 12 08:14:46 crc kubenswrapper[4867]: I1212 08:14:46.851548 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" path="/var/lib/kubelet/pods/a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d/volumes" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.560167 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:14:47 crc kubenswrapper[4867]: E1212 08:14:47.561242 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="registry-server" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.561266 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="registry-server" Dec 12 08:14:47 crc kubenswrapper[4867]: E1212 08:14:47.561280 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="extract-utilities" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.561287 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="extract-utilities" Dec 12 08:14:47 crc kubenswrapper[4867]: E1212 08:14:47.561318 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="extract-content" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.561324 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="extract-content" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.561520 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9c34d4c-4c95-46f5-875a-6c47fc2a1e2d" containerName="registry-server" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.562893 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.585495 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.737646 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.738519 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kl8k\" (UniqueName: \"kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.738699 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.840565 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.840885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kl8k\" (UniqueName: \"kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.840981 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.841527 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.841831 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.860892 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kl8k\" (UniqueName: \"kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k\") pod \"redhat-marketplace-bkf42\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:47 crc kubenswrapper[4867]: I1212 08:14:47.883567 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:48 crc kubenswrapper[4867]: I1212 08:14:48.406765 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:14:49 crc kubenswrapper[4867]: I1212 08:14:49.428774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerStarted","Data":"011d39b7be5050e04d1357c5acdba137b927c164e6cda241b998caae3e8faff9"} Dec 12 08:14:50 crc kubenswrapper[4867]: I1212 08:14:50.438840 4867 generic.go:334] "Generic (PLEG): container finished" podID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerID="8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e" exitCode=0 Dec 12 08:14:50 crc kubenswrapper[4867]: I1212 08:14:50.438940 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerDied","Data":"8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e"} Dec 12 08:14:52 crc kubenswrapper[4867]: I1212 08:14:52.457295 4867 generic.go:334] "Generic (PLEG): container finished" podID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerID="260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f" exitCode=0 Dec 12 08:14:52 crc kubenswrapper[4867]: I1212 08:14:52.457369 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerDied","Data":"260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f"} Dec 12 08:14:54 crc kubenswrapper[4867]: I1212 08:14:54.473731 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerStarted","Data":"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26"} Dec 12 08:14:54 crc kubenswrapper[4867]: I1212 08:14:54.494660 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bkf42" podStartSLOduration=4.124405314 podStartE2EDuration="7.494638595s" podCreationTimestamp="2025-12-12 08:14:47 +0000 UTC" firstStartedPulling="2025-12-12 08:14:50.440401654 +0000 UTC m=+5178.011782923" lastFinishedPulling="2025-12-12 08:14:53.810634935 +0000 UTC m=+5181.382016204" observedRunningTime="2025-12-12 08:14:54.49241483 +0000 UTC m=+5182.063796099" watchObservedRunningTime="2025-12-12 08:14:54.494638595 +0000 UTC m=+5182.066019864" Dec 12 08:14:57 crc kubenswrapper[4867]: I1212 08:14:57.883875 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:57 crc kubenswrapper[4867]: I1212 08:14:57.884386 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:57 crc kubenswrapper[4867]: I1212 08:14:57.932644 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:58 crc kubenswrapper[4867]: I1212 08:14:58.553757 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:14:58 crc kubenswrapper[4867]: I1212 08:14:58.602539 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.140003 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v"] Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.143886 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.146212 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.147656 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.149868 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.149943 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcztl\" (UniqueName: \"kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.150034 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.170257 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v"] Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.252173 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.253324 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcztl\" (UniqueName: \"kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.253902 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.254771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.258631 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.271353 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcztl\" (UniqueName: \"kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl\") pod \"collect-profiles-29425455-nwl8v\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.463177 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.529967 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bkf42" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="registry-server" containerID="cri-o://902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26" gracePeriod=2 Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.904832 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:15:00 crc kubenswrapper[4867]: I1212 08:15:00.948000 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v"] Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.068323 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities\") pod \"c900e3f0-62c2-4fdd-a496-df32f76f3851\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.068985 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content\") pod \"c900e3f0-62c2-4fdd-a496-df32f76f3851\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.069186 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kl8k\" (UniqueName: \"kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k\") pod \"c900e3f0-62c2-4fdd-a496-df32f76f3851\" (UID: \"c900e3f0-62c2-4fdd-a496-df32f76f3851\") " Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.069709 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities" (OuterVolumeSpecName: "utilities") pod "c900e3f0-62c2-4fdd-a496-df32f76f3851" (UID: "c900e3f0-62c2-4fdd-a496-df32f76f3851"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.070441 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.075044 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k" (OuterVolumeSpecName: "kube-api-access-9kl8k") pod "c900e3f0-62c2-4fdd-a496-df32f76f3851" (UID: "c900e3f0-62c2-4fdd-a496-df32f76f3851"). InnerVolumeSpecName "kube-api-access-9kl8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.096709 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c900e3f0-62c2-4fdd-a496-df32f76f3851" (UID: "c900e3f0-62c2-4fdd-a496-df32f76f3851"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.171650 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c900e3f0-62c2-4fdd-a496-df32f76f3851-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.171687 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kl8k\" (UniqueName: \"kubernetes.io/projected/c900e3f0-62c2-4fdd-a496-df32f76f3851-kube-api-access-9kl8k\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.539409 4867 generic.go:334] "Generic (PLEG): container finished" podID="28114eb1-01be-4a03-a4db-60f1848c3d05" containerID="69d2dc93878f543632eb1060dad7aedfcfe1ec67c2b8cbee6a5779f1dc007983" exitCode=0 Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.539448 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" event={"ID":"28114eb1-01be-4a03-a4db-60f1848c3d05","Type":"ContainerDied","Data":"69d2dc93878f543632eb1060dad7aedfcfe1ec67c2b8cbee6a5779f1dc007983"} Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.539501 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" event={"ID":"28114eb1-01be-4a03-a4db-60f1848c3d05","Type":"ContainerStarted","Data":"8626f37006e4b9826740018c04825e3fab538d37288eff13e100e6bed0174b4d"} Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.542237 4867 generic.go:334] "Generic (PLEG): container finished" podID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerID="902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26" exitCode=0 Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.542288 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerDied","Data":"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26"} Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.542339 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bkf42" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.542355 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bkf42" event={"ID":"c900e3f0-62c2-4fdd-a496-df32f76f3851","Type":"ContainerDied","Data":"011d39b7be5050e04d1357c5acdba137b927c164e6cda241b998caae3e8faff9"} Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.542378 4867 scope.go:117] "RemoveContainer" containerID="902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.563648 4867 scope.go:117] "RemoveContainer" containerID="260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.572498 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.579047 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bkf42"] Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.602736 4867 scope.go:117] "RemoveContainer" containerID="8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.635344 4867 scope.go:117] "RemoveContainer" containerID="902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26" Dec 12 08:15:01 crc kubenswrapper[4867]: E1212 08:15:01.636175 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26\": container with ID starting with 902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26 not found: ID does not exist" containerID="902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.636245 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26"} err="failed to get container status \"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26\": rpc error: code = NotFound desc = could not find container \"902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26\": container with ID starting with 902435856e3856fb195cba8724780e1b5bc8572021553c4bf4302e6a7b83cf26 not found: ID does not exist" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.636283 4867 scope.go:117] "RemoveContainer" containerID="260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f" Dec 12 08:15:01 crc kubenswrapper[4867]: E1212 08:15:01.636743 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f\": container with ID starting with 260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f not found: ID does not exist" containerID="260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.636772 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f"} err="failed to get container status \"260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f\": rpc error: code = NotFound desc = could not find container \"260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f\": container with ID starting with 260bd4cffbcad7ab4a12ee0c2c9080b551aa8a783571b6c28a332919ed1c603f not found: ID does not exist" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.636790 4867 scope.go:117] "RemoveContainer" containerID="8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e" Dec 12 08:15:01 crc kubenswrapper[4867]: E1212 08:15:01.637322 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e\": container with ID starting with 8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e not found: ID does not exist" containerID="8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e" Dec 12 08:15:01 crc kubenswrapper[4867]: I1212 08:15:01.637355 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e"} err="failed to get container status \"8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e\": rpc error: code = NotFound desc = could not find container \"8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e\": container with ID starting with 8e0a375292725d15914c4a51529bd0333b7637a6eccc9a3ea5cdf62f8008cd5e not found: ID does not exist" Dec 12 08:15:02 crc kubenswrapper[4867]: I1212 08:15:02.854402 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" path="/var/lib/kubelet/pods/c900e3f0-62c2-4fdd-a496-df32f76f3851/volumes" Dec 12 08:15:02 crc kubenswrapper[4867]: I1212 08:15:02.864722 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.000473 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume\") pod \"28114eb1-01be-4a03-a4db-60f1848c3d05\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.000531 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcztl\" (UniqueName: \"kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl\") pod \"28114eb1-01be-4a03-a4db-60f1848c3d05\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.000574 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume\") pod \"28114eb1-01be-4a03-a4db-60f1848c3d05\" (UID: \"28114eb1-01be-4a03-a4db-60f1848c3d05\") " Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.001787 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume" (OuterVolumeSpecName: "config-volume") pod "28114eb1-01be-4a03-a4db-60f1848c3d05" (UID: "28114eb1-01be-4a03-a4db-60f1848c3d05"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.009135 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl" (OuterVolumeSpecName: "kube-api-access-zcztl") pod "28114eb1-01be-4a03-a4db-60f1848c3d05" (UID: "28114eb1-01be-4a03-a4db-60f1848c3d05"). InnerVolumeSpecName "kube-api-access-zcztl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.009350 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "28114eb1-01be-4a03-a4db-60f1848c3d05" (UID: "28114eb1-01be-4a03-a4db-60f1848c3d05"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.102396 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/28114eb1-01be-4a03-a4db-60f1848c3d05-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.102426 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcztl\" (UniqueName: \"kubernetes.io/projected/28114eb1-01be-4a03-a4db-60f1848c3d05-kube-api-access-zcztl\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.102435 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/28114eb1-01be-4a03-a4db-60f1848c3d05-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.561608 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" event={"ID":"28114eb1-01be-4a03-a4db-60f1848c3d05","Type":"ContainerDied","Data":"8626f37006e4b9826740018c04825e3fab538d37288eff13e100e6bed0174b4d"} Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.561649 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8626f37006e4b9826740018c04825e3fab538d37288eff13e100e6bed0174b4d" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.561699 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v" Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.936783 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc"] Dec 12 08:15:03 crc kubenswrapper[4867]: I1212 08:15:03.946481 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425410-tpdxc"] Dec 12 08:15:04 crc kubenswrapper[4867]: I1212 08:15:04.852191 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79b364c-f390-4159-a7a4-25f976acdd87" path="/var/lib/kubelet/pods/b79b364c-f390-4159-a7a4-25f976acdd87/volumes" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.145304 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:14 crc kubenswrapper[4867]: E1212 08:15:14.147391 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="extract-utilities" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147510 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="extract-utilities" Dec 12 08:15:14 crc kubenswrapper[4867]: E1212 08:15:14.147621 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="registry-server" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147643 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="registry-server" Dec 12 08:15:14 crc kubenswrapper[4867]: E1212 08:15:14.147660 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28114eb1-01be-4a03-a4db-60f1848c3d05" containerName="collect-profiles" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147669 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="28114eb1-01be-4a03-a4db-60f1848c3d05" containerName="collect-profiles" Dec 12 08:15:14 crc kubenswrapper[4867]: E1212 08:15:14.147687 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="extract-content" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147693 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="extract-content" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147883 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="28114eb1-01be-4a03-a4db-60f1848c3d05" containerName="collect-profiles" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.147895 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c900e3f0-62c2-4fdd-a496-df32f76f3851" containerName="registry-server" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.149035 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.160984 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.291507 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.291548 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.291799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcbvb\" (UniqueName: \"kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.393378 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.393427 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.393550 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcbvb\" (UniqueName: \"kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.393996 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.394189 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.416335 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcbvb\" (UniqueName: \"kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb\") pod \"community-operators-mvh6s\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:14 crc kubenswrapper[4867]: I1212 08:15:14.473633 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:15 crc kubenswrapper[4867]: I1212 08:15:15.006496 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:15 crc kubenswrapper[4867]: I1212 08:15:15.663938 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c07f042-2371-4035-8530-448dfce60043" containerID="e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a" exitCode=0 Dec 12 08:15:15 crc kubenswrapper[4867]: I1212 08:15:15.663987 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerDied","Data":"e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a"} Dec 12 08:15:15 crc kubenswrapper[4867]: I1212 08:15:15.664032 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerStarted","Data":"307337fe8600f77138769298ed43f6975e53abb284d9dc8801daf305d43320eb"} Dec 12 08:15:16 crc kubenswrapper[4867]: I1212 08:15:16.682922 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerStarted","Data":"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7"} Dec 12 08:15:17 crc kubenswrapper[4867]: I1212 08:15:17.693659 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c07f042-2371-4035-8530-448dfce60043" containerID="21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7" exitCode=0 Dec 12 08:15:17 crc kubenswrapper[4867]: I1212 08:15:17.693704 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerDied","Data":"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7"} Dec 12 08:15:18 crc kubenswrapper[4867]: I1212 08:15:18.721049 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerStarted","Data":"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b"} Dec 12 08:15:18 crc kubenswrapper[4867]: I1212 08:15:18.746897 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mvh6s" podStartSLOduration=2.244209864 podStartE2EDuration="4.746874404s" podCreationTimestamp="2025-12-12 08:15:14 +0000 UTC" firstStartedPulling="2025-12-12 08:15:15.665529769 +0000 UTC m=+5203.236911038" lastFinishedPulling="2025-12-12 08:15:18.168194309 +0000 UTC m=+5205.739575578" observedRunningTime="2025-12-12 08:15:18.742290851 +0000 UTC m=+5206.313672140" watchObservedRunningTime="2025-12-12 08:15:18.746874404 +0000 UTC m=+5206.318255683" Dec 12 08:15:24 crc kubenswrapper[4867]: I1212 08:15:24.474417 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:24 crc kubenswrapper[4867]: I1212 08:15:24.475274 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:24 crc kubenswrapper[4867]: I1212 08:15:24.523952 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:24 crc kubenswrapper[4867]: I1212 08:15:24.809384 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:24 crc kubenswrapper[4867]: I1212 08:15:24.858548 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:26 crc kubenswrapper[4867]: I1212 08:15:26.782290 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mvh6s" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="registry-server" containerID="cri-o://67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b" gracePeriod=2 Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.271506 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.411117 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities\") pod \"8c07f042-2371-4035-8530-448dfce60043\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.411265 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content\") pod \"8c07f042-2371-4035-8530-448dfce60043\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.411317 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcbvb\" (UniqueName: \"kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb\") pod \"8c07f042-2371-4035-8530-448dfce60043\" (UID: \"8c07f042-2371-4035-8530-448dfce60043\") " Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.412736 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities" (OuterVolumeSpecName: "utilities") pod "8c07f042-2371-4035-8530-448dfce60043" (UID: "8c07f042-2371-4035-8530-448dfce60043"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.416313 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb" (OuterVolumeSpecName: "kube-api-access-bcbvb") pod "8c07f042-2371-4035-8530-448dfce60043" (UID: "8c07f042-2371-4035-8530-448dfce60043"). InnerVolumeSpecName "kube-api-access-bcbvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.459501 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c07f042-2371-4035-8530-448dfce60043" (UID: "8c07f042-2371-4035-8530-448dfce60043"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.513288 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.513328 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c07f042-2371-4035-8530-448dfce60043-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.513341 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcbvb\" (UniqueName: \"kubernetes.io/projected/8c07f042-2371-4035-8530-448dfce60043-kube-api-access-bcbvb\") on node \"crc\" DevicePath \"\"" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.794456 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c07f042-2371-4035-8530-448dfce60043" containerID="67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b" exitCode=0 Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.794561 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerDied","Data":"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b"} Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.794628 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvh6s" event={"ID":"8c07f042-2371-4035-8530-448dfce60043","Type":"ContainerDied","Data":"307337fe8600f77138769298ed43f6975e53abb284d9dc8801daf305d43320eb"} Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.794684 4867 scope.go:117] "RemoveContainer" containerID="67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.795254 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvh6s" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.818215 4867 scope.go:117] "RemoveContainer" containerID="21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.848395 4867 scope.go:117] "RemoveContainer" containerID="e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.890173 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.891360 4867 scope.go:117] "RemoveContainer" containerID="67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b" Dec 12 08:15:27 crc kubenswrapper[4867]: E1212 08:15:27.892032 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b\": container with ID starting with 67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b not found: ID does not exist" containerID="67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.892067 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b"} err="failed to get container status \"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b\": rpc error: code = NotFound desc = could not find container \"67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b\": container with ID starting with 67def49daf7478f7d57fef6ca93910fe56641115d22fb14a0c7cdcb8479d955b not found: ID does not exist" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.892092 4867 scope.go:117] "RemoveContainer" containerID="21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7" Dec 12 08:15:27 crc kubenswrapper[4867]: E1212 08:15:27.892560 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7\": container with ID starting with 21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7 not found: ID does not exist" containerID="21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.892610 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7"} err="failed to get container status \"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7\": rpc error: code = NotFound desc = could not find container \"21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7\": container with ID starting with 21b3405fab34e6eb43a40218e82b9d464d6d048b3399050cebe7da27172dfba7 not found: ID does not exist" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.892728 4867 scope.go:117] "RemoveContainer" containerID="e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a" Dec 12 08:15:27 crc kubenswrapper[4867]: E1212 08:15:27.893068 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a\": container with ID starting with e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a not found: ID does not exist" containerID="e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.893091 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a"} err="failed to get container status \"e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a\": rpc error: code = NotFound desc = could not find container \"e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a\": container with ID starting with e56571e4b667fac70f8577cbdaaf6f7a45f7e52faf90badf9dc3d4ccd7256f3a not found: ID does not exist" Dec 12 08:15:27 crc kubenswrapper[4867]: I1212 08:15:27.897716 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mvh6s"] Dec 12 08:15:28 crc kubenswrapper[4867]: I1212 08:15:28.850293 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c07f042-2371-4035-8530-448dfce60043" path="/var/lib/kubelet/pods/8c07f042-2371-4035-8530-448dfce60043/volumes" Dec 12 08:15:28 crc kubenswrapper[4867]: I1212 08:15:28.989573 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:15:28 crc kubenswrapper[4867]: I1212 08:15:28.989627 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:15:45 crc kubenswrapper[4867]: I1212 08:15:45.811690 4867 scope.go:117] "RemoveContainer" containerID="6b93c9252fbdbe34515b9e77e48416d5b8a91f09b403c5f5f9f613bf0febc14f" Dec 12 08:15:45 crc kubenswrapper[4867]: I1212 08:15:45.853613 4867 scope.go:117] "RemoveContainer" containerID="5ab635b37604298b1c5cfaee2df09d922f07070c79a48401bc5f0151a08a58dd" Dec 12 08:15:45 crc kubenswrapper[4867]: I1212 08:15:45.880618 4867 scope.go:117] "RemoveContainer" containerID="8ed629835915dc25b9d0d93c36970c9468816c30804aa7e8fa1c1c8680cde4b2" Dec 12 08:15:45 crc kubenswrapper[4867]: I1212 08:15:45.917974 4867 scope.go:117] "RemoveContainer" containerID="22bfbc2c8546c31b207544221b6aff44ae17b65b2f1a8a60b90ab479287b0729" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.614193 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-2mpzh"] Dec 12 08:15:54 crc kubenswrapper[4867]: E1212 08:15:54.615059 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="registry-server" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.615073 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="registry-server" Dec 12 08:15:54 crc kubenswrapper[4867]: E1212 08:15:54.615089 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="extract-content" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.615095 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="extract-content" Dec 12 08:15:54 crc kubenswrapper[4867]: E1212 08:15:54.615113 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="extract-utilities" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.615120 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="extract-utilities" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.615312 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c07f042-2371-4035-8530-448dfce60043" containerName="registry-server" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.615875 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.619446 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f6pb\" (UniqueName: \"kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.619623 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.627003 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2mpzh"] Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.634550 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8af0-account-create-update-nmdss"] Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.635652 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.637782 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.648281 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8af0-account-create-update-nmdss"] Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.721008 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.721097 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.721153 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f6pb\" (UniqueName: \"kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.721291 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-766h7\" (UniqueName: \"kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.722080 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.763150 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f6pb\" (UniqueName: \"kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb\") pod \"barbican-db-create-2mpzh\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.822188 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-766h7\" (UniqueName: \"kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.822320 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.823247 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.857742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-766h7\" (UniqueName: \"kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7\") pod \"barbican-8af0-account-create-update-nmdss\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.936248 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2mpzh" Dec 12 08:15:54 crc kubenswrapper[4867]: I1212 08:15:54.956854 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:15:55 crc kubenswrapper[4867]: I1212 08:15:55.425033 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-2mpzh"] Dec 12 08:15:55 crc kubenswrapper[4867]: I1212 08:15:55.497217 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8af0-account-create-update-nmdss"] Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.066521 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8af0-account-create-update-nmdss" event={"ID":"9212e1c2-e6d1-463a-b33b-323d72d37fec","Type":"ContainerStarted","Data":"5dfad9aee108d994e453cf359b424a8cb7c3d9aa73fb060ecb090f17b311b9c1"} Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.066566 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8af0-account-create-update-nmdss" event={"ID":"9212e1c2-e6d1-463a-b33b-323d72d37fec","Type":"ContainerStarted","Data":"57b9eba27b251d5e2a4e7508e7acdfaf0bb6a69e4f4748c9a81a897988b0b75f"} Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.068739 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2mpzh" event={"ID":"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d","Type":"ContainerStarted","Data":"1b46710b606b3e148a086b597a61d7e794f4ce1500c5ba0aa27b018ef8624166"} Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.068774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2mpzh" event={"ID":"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d","Type":"ContainerStarted","Data":"1ddf4e8a902a0a90d5204472ef7aaff52938b090ed6d7b7f8e62ce51c431797e"} Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.085334 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-8af0-account-create-update-nmdss" podStartSLOduration=2.08531752 podStartE2EDuration="2.08531752s" podCreationTimestamp="2025-12-12 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:15:56.083772561 +0000 UTC m=+5243.655153830" watchObservedRunningTime="2025-12-12 08:15:56.08531752 +0000 UTC m=+5243.656698789" Dec 12 08:15:56 crc kubenswrapper[4867]: I1212 08:15:56.100522 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-2mpzh" podStartSLOduration=2.100504484 podStartE2EDuration="2.100504484s" podCreationTimestamp="2025-12-12 08:15:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:15:56.098493144 +0000 UTC m=+5243.669874413" watchObservedRunningTime="2025-12-12 08:15:56.100504484 +0000 UTC m=+5243.671885753" Dec 12 08:15:58 crc kubenswrapper[4867]: I1212 08:15:58.988546 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:15:58 crc kubenswrapper[4867]: I1212 08:15:58.989092 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:16:00 crc kubenswrapper[4867]: I1212 08:16:00.105682 4867 generic.go:334] "Generic (PLEG): container finished" podID="5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" containerID="1b46710b606b3e148a086b597a61d7e794f4ce1500c5ba0aa27b018ef8624166" exitCode=0 Dec 12 08:16:00 crc kubenswrapper[4867]: I1212 08:16:00.105787 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2mpzh" event={"ID":"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d","Type":"ContainerDied","Data":"1b46710b606b3e148a086b597a61d7e794f4ce1500c5ba0aa27b018ef8624166"} Dec 12 08:16:00 crc kubenswrapper[4867]: I1212 08:16:00.107944 4867 generic.go:334] "Generic (PLEG): container finished" podID="9212e1c2-e6d1-463a-b33b-323d72d37fec" containerID="5dfad9aee108d994e453cf359b424a8cb7c3d9aa73fb060ecb090f17b311b9c1" exitCode=0 Dec 12 08:16:00 crc kubenswrapper[4867]: I1212 08:16:00.107993 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8af0-account-create-update-nmdss" event={"ID":"9212e1c2-e6d1-463a-b33b-323d72d37fec","Type":"ContainerDied","Data":"5dfad9aee108d994e453cf359b424a8cb7c3d9aa73fb060ecb090f17b311b9c1"} Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.501084 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.516810 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2mpzh" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.646536 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts\") pod \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.646788 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4f6pb\" (UniqueName: \"kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb\") pod \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\" (UID: \"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d\") " Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.646810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-766h7\" (UniqueName: \"kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7\") pod \"9212e1c2-e6d1-463a-b33b-323d72d37fec\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.646896 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts\") pod \"9212e1c2-e6d1-463a-b33b-323d72d37fec\" (UID: \"9212e1c2-e6d1-463a-b33b-323d72d37fec\") " Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.647298 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" (UID: "5764dd52-d7f4-483f-93c9-fd5a70ef6a1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.647752 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9212e1c2-e6d1-463a-b33b-323d72d37fec" (UID: "9212e1c2-e6d1-463a-b33b-323d72d37fec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.652817 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb" (OuterVolumeSpecName: "kube-api-access-4f6pb") pod "5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" (UID: "5764dd52-d7f4-483f-93c9-fd5a70ef6a1d"). InnerVolumeSpecName "kube-api-access-4f6pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.653335 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7" (OuterVolumeSpecName: "kube-api-access-766h7") pod "9212e1c2-e6d1-463a-b33b-323d72d37fec" (UID: "9212e1c2-e6d1-463a-b33b-323d72d37fec"). InnerVolumeSpecName "kube-api-access-766h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.748745 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9212e1c2-e6d1-463a-b33b-323d72d37fec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.749032 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.749279 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4f6pb\" (UniqueName: \"kubernetes.io/projected/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d-kube-api-access-4f6pb\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:01 crc kubenswrapper[4867]: I1212 08:16:01.749371 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-766h7\" (UniqueName: \"kubernetes.io/projected/9212e1c2-e6d1-463a-b33b-323d72d37fec-kube-api-access-766h7\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.125407 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8af0-account-create-update-nmdss" event={"ID":"9212e1c2-e6d1-463a-b33b-323d72d37fec","Type":"ContainerDied","Data":"57b9eba27b251d5e2a4e7508e7acdfaf0bb6a69e4f4748c9a81a897988b0b75f"} Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.125766 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57b9eba27b251d5e2a4e7508e7acdfaf0bb6a69e4f4748c9a81a897988b0b75f" Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.125425 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8af0-account-create-update-nmdss" Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.127084 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-2mpzh" event={"ID":"5764dd52-d7f4-483f-93c9-fd5a70ef6a1d","Type":"ContainerDied","Data":"1ddf4e8a902a0a90d5204472ef7aaff52938b090ed6d7b7f8e62ce51c431797e"} Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.127110 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ddf4e8a902a0a90d5204472ef7aaff52938b090ed6d7b7f8e62ce51c431797e" Dec 12 08:16:02 crc kubenswrapper[4867]: I1212 08:16:02.127178 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-2mpzh" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.917022 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2rzp2"] Dec 12 08:16:04 crc kubenswrapper[4867]: E1212 08:16:04.917691 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" containerName="mariadb-database-create" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.917706 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" containerName="mariadb-database-create" Dec 12 08:16:04 crc kubenswrapper[4867]: E1212 08:16:04.917727 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9212e1c2-e6d1-463a-b33b-323d72d37fec" containerName="mariadb-account-create-update" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.917732 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9212e1c2-e6d1-463a-b33b-323d72d37fec" containerName="mariadb-account-create-update" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.917915 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" containerName="mariadb-database-create" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.917944 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9212e1c2-e6d1-463a-b33b-323d72d37fec" containerName="mariadb-account-create-update" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.918494 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.921485 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-wjlzx" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.921511 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 12 08:16:04 crc kubenswrapper[4867]: I1212 08:16:04.927439 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2rzp2"] Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.103746 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwjl\" (UniqueName: \"kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.103889 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.104040 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.205055 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.205128 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwjl\" (UniqueName: \"kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.205180 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.215074 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.215646 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.222472 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwjl\" (UniqueName: \"kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl\") pod \"barbican-db-sync-2rzp2\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.250423 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:05 crc kubenswrapper[4867]: I1212 08:16:05.742450 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2rzp2"] Dec 12 08:16:06 crc kubenswrapper[4867]: I1212 08:16:06.186484 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2rzp2" event={"ID":"71445001-629a-4b2c-a78f-428aa0164958","Type":"ContainerStarted","Data":"4e9f4e5a963764cf82fec96603433010dae049164f30be34182493b9845e05fa"} Dec 12 08:16:12 crc kubenswrapper[4867]: I1212 08:16:12.229477 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2rzp2" event={"ID":"71445001-629a-4b2c-a78f-428aa0164958","Type":"ContainerStarted","Data":"0d90299f10387afb4d7c2724416a7468daf60a136d919a64c99eb93ffc0fa997"} Dec 12 08:16:12 crc kubenswrapper[4867]: I1212 08:16:12.250166 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2rzp2" podStartSLOduration=2.205051237 podStartE2EDuration="8.25014389s" podCreationTimestamp="2025-12-12 08:16:04 +0000 UTC" firstStartedPulling="2025-12-12 08:16:05.759005653 +0000 UTC m=+5253.330386922" lastFinishedPulling="2025-12-12 08:16:11.804098306 +0000 UTC m=+5259.375479575" observedRunningTime="2025-12-12 08:16:12.245889775 +0000 UTC m=+5259.817271044" watchObservedRunningTime="2025-12-12 08:16:12.25014389 +0000 UTC m=+5259.821525159" Dec 12 08:16:14 crc kubenswrapper[4867]: I1212 08:16:14.252974 4867 generic.go:334] "Generic (PLEG): container finished" podID="71445001-629a-4b2c-a78f-428aa0164958" containerID="0d90299f10387afb4d7c2724416a7468daf60a136d919a64c99eb93ffc0fa997" exitCode=0 Dec 12 08:16:14 crc kubenswrapper[4867]: I1212 08:16:14.253419 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2rzp2" event={"ID":"71445001-629a-4b2c-a78f-428aa0164958","Type":"ContainerDied","Data":"0d90299f10387afb4d7c2724416a7468daf60a136d919a64c99eb93ffc0fa997"} Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.658374 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.780911 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle\") pod \"71445001-629a-4b2c-a78f-428aa0164958\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.780977 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data\") pod \"71445001-629a-4b2c-a78f-428aa0164958\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.781093 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hwjl\" (UniqueName: \"kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl\") pod \"71445001-629a-4b2c-a78f-428aa0164958\" (UID: \"71445001-629a-4b2c-a78f-428aa0164958\") " Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.787277 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "71445001-629a-4b2c-a78f-428aa0164958" (UID: "71445001-629a-4b2c-a78f-428aa0164958"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.787426 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl" (OuterVolumeSpecName: "kube-api-access-6hwjl") pod "71445001-629a-4b2c-a78f-428aa0164958" (UID: "71445001-629a-4b2c-a78f-428aa0164958"). InnerVolumeSpecName "kube-api-access-6hwjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.819730 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71445001-629a-4b2c-a78f-428aa0164958" (UID: "71445001-629a-4b2c-a78f-428aa0164958"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.883653 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.883970 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/71445001-629a-4b2c-a78f-428aa0164958-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:15 crc kubenswrapper[4867]: I1212 08:16:15.883982 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hwjl\" (UniqueName: \"kubernetes.io/projected/71445001-629a-4b2c-a78f-428aa0164958-kube-api-access-6hwjl\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.284004 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2rzp2" event={"ID":"71445001-629a-4b2c-a78f-428aa0164958","Type":"ContainerDied","Data":"4e9f4e5a963764cf82fec96603433010dae049164f30be34182493b9845e05fa"} Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.284046 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9f4e5a963764cf82fec96603433010dae049164f30be34182493b9845e05fa" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.284051 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2rzp2" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.499104 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84b67f6bf7-jqpkj"] Dec 12 08:16:16 crc kubenswrapper[4867]: E1212 08:16:16.499495 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71445001-629a-4b2c-a78f-428aa0164958" containerName="barbican-db-sync" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.499512 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="71445001-629a-4b2c-a78f-428aa0164958" containerName="barbican-db-sync" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.499662 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="71445001-629a-4b2c-a78f-428aa0164958" containerName="barbican-db-sync" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.500571 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.504925 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-wjlzx" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.505420 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.508163 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.524378 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84b67f6bf7-jqpkj"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.531510 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5bff489498-h54sr"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.532808 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.540652 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.564111 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bff489498-h54sr"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.599446 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data-custom\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.599572 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.599606 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ad1843a-0c00-40d2-95b5-4f164cc54053-logs\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.599653 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5g49\" (UniqueName: \"kubernetes.io/projected/2ad1843a-0c00-40d2-95b5-4f164cc54053-kube-api-access-z5g49\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.599717 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-combined-ca-bundle\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.635391 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.637134 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.660681 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703168 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5g49\" (UniqueName: \"kubernetes.io/projected/2ad1843a-0c00-40d2-95b5-4f164cc54053-kube-api-access-z5g49\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703560 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e41196ac-a7c9-47fa-9ca0-584a5fad129b-logs\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703615 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-combined-ca-bundle\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703653 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data-custom\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703700 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data-custom\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703738 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bxtp\" (UniqueName: \"kubernetes.io/projected/e41196ac-a7c9-47fa-9ca0-584a5fad129b-kube-api-access-8bxtp\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703774 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703796 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-combined-ca-bundle\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703836 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.703858 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ad1843a-0c00-40d2-95b5-4f164cc54053-logs\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.704326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ad1843a-0c00-40d2-95b5-4f164cc54053-logs\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.706370 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.708074 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.711441 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.716197 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.718618 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data-custom\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.724473 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-config-data\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.743813 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad1843a-0c00-40d2-95b5-4f164cc54053-combined-ca-bundle\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.757892 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5g49\" (UniqueName: \"kubernetes.io/projected/2ad1843a-0c00-40d2-95b5-4f164cc54053-kube-api-access-z5g49\") pod \"barbican-worker-84b67f6bf7-jqpkj\" (UID: \"2ad1843a-0c00-40d2-95b5-4f164cc54053\") " pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.805800 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.805862 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806358 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e41196ac-a7c9-47fa-9ca0-584a5fad129b-logs\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806408 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806432 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806463 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806513 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806547 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data-custom\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806571 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806612 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bxtp\" (UniqueName: \"kubernetes.io/projected/e41196ac-a7c9-47fa-9ca0-584a5fad129b-kube-api-access-8bxtp\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806689 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wnj\" (UniqueName: \"kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806719 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806760 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806776 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-combined-ca-bundle\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.806795 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd84j\" (UniqueName: \"kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.807625 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e41196ac-a7c9-47fa-9ca0-584a5fad129b-logs\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.813135 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-combined-ca-bundle\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.813867 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data-custom\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.814512 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e41196ac-a7c9-47fa-9ca0-584a5fad129b-config-data\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.826628 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.827139 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bxtp\" (UniqueName: \"kubernetes.io/projected/e41196ac-a7c9-47fa-9ca0-584a5fad129b-kube-api-access-8bxtp\") pod \"barbican-keystone-listener-5bff489498-h54sr\" (UID: \"e41196ac-a7c9-47fa-9ca0-584a5fad129b\") " pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.860620 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.908589 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.909883 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.910482 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.910520 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.910642 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.911450 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.912095 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.912626 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wnj\" (UniqueName: \"kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.913083 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.913116 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd84j\" (UniqueName: \"kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.913177 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.913210 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.913654 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.915194 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.916300 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.916655 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.917476 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.917868 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.941567 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wnj\" (UniqueName: \"kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj\") pod \"barbican-api-74f48cfc74-rblds\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.943814 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd84j\" (UniqueName: \"kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j\") pod \"dnsmasq-dns-79dc9d59f9-7jslg\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.949960 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.965416 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:16 crc kubenswrapper[4867]: I1212 08:16:16.977152 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:16.999350 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.002007 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.119204 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.119323 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.119434 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7q66\" (UniqueName: \"kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.223177 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7q66\" (UniqueName: \"kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.223578 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.223642 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.224260 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.225883 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.251218 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7q66\" (UniqueName: \"kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66\") pod \"redhat-operators-sdvfn\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.340013 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5bff489498-h54sr"] Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.342194 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.431261 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84b67f6bf7-jqpkj"] Dec 12 08:16:17 crc kubenswrapper[4867]: W1212 08:16:17.450819 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ad1843a_0c00_40d2_95b5_4f164cc54053.slice/crio-4f153d2b6e9e830715f80c94090abdfbe637d1ed6c932f8313723f9b09fcb8ed WatchSource:0}: Error finding container 4f153d2b6e9e830715f80c94090abdfbe637d1ed6c932f8313723f9b09fcb8ed: Status 404 returned error can't find the container with id 4f153d2b6e9e830715f80c94090abdfbe637d1ed6c932f8313723f9b09fcb8ed Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.707448 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.727878 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:17 crc kubenswrapper[4867]: W1212 08:16:17.742715 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4345a89_22d4_4cb2_a6d2_dcb4efe1e1d2.slice/crio-77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b WatchSource:0}: Error finding container 77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b: Status 404 returned error can't find the container with id 77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b Dec 12 08:16:17 crc kubenswrapper[4867]: I1212 08:16:17.761431 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.339628 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerStarted","Data":"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.339891 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerStarted","Data":"bbe3e8aedf23d098d972a074b1a520afc812a44ee16dd09c50b64d23bec416c8"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.354915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" event={"ID":"e41196ac-a7c9-47fa-9ca0-584a5fad129b","Type":"ContainerStarted","Data":"f4cfabb3c737e2eff21a65de14fa892acda02b5270aaf18ddf89ddbf752ba5f2"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.369401 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" event={"ID":"2ad1843a-0c00-40d2-95b5-4f164cc54053","Type":"ContainerStarted","Data":"4f153d2b6e9e830715f80c94090abdfbe637d1ed6c932f8313723f9b09fcb8ed"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.388718 4867 generic.go:334] "Generic (PLEG): container finished" podID="1c579316-8df2-4a9f-a386-c2c97b101676" containerID="1688f4f05e2f518d285a162d40cc81d96ec60b5b606dbbc05c8ee927c834405c" exitCode=0 Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.388833 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerDied","Data":"1688f4f05e2f518d285a162d40cc81d96ec60b5b606dbbc05c8ee927c834405c"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.388870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerStarted","Data":"a37b75bddca640ed3f304f24f4b0a327f9961f2f66ebd33a5986e41bafa828dd"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.415539 4867 generic.go:334] "Generic (PLEG): container finished" podID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerID="b70bca5053c6d4ed48d8a30e799ddbd6632b5b3e4e0f73f8f3fd8c01accc596c" exitCode=0 Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.415592 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" event={"ID":"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2","Type":"ContainerDied","Data":"b70bca5053c6d4ed48d8a30e799ddbd6632b5b3e4e0f73f8f3fd8c01accc596c"} Dec 12 08:16:18 crc kubenswrapper[4867]: I1212 08:16:18.415640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" event={"ID":"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2","Type":"ContainerStarted","Data":"77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b"} Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.438422 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" event={"ID":"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2","Type":"ContainerStarted","Data":"dedf6ca4ad16f45eb6bc92f2324981d45457f8ecc3f587f476394ca47e89cb58"} Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.440118 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.445740 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerStarted","Data":"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253"} Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.445897 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.445934 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.465272 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" podStartSLOduration=3.4652516159999998 podStartE2EDuration="3.465251616s" podCreationTimestamp="2025-12-12 08:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:16:19.458607893 +0000 UTC m=+5267.029989162" watchObservedRunningTime="2025-12-12 08:16:19.465251616 +0000 UTC m=+5267.036632885" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.485914 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74f48cfc74-rblds" podStartSLOduration=3.485898794 podStartE2EDuration="3.485898794s" podCreationTimestamp="2025-12-12 08:16:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:16:19.48495323 +0000 UTC m=+5267.056334499" watchObservedRunningTime="2025-12-12 08:16:19.485898794 +0000 UTC m=+5267.057280063" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.564962 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cbd859876-xpw6c"] Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.566502 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.569632 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.571504 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.582984 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cbd859876-xpw6c"] Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704612 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data-custom\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704728 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-combined-ca-bundle\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704772 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a54eb8d-8507-4f32-b2bf-936ea487c701-logs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704792 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704881 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5997\" (UniqueName: \"kubernetes.io/projected/7a54eb8d-8507-4f32-b2bf-936ea487c701-kube-api-access-d5997\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704925 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-public-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.704969 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-internal-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807029 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5997\" (UniqueName: \"kubernetes.io/projected/7a54eb8d-8507-4f32-b2bf-936ea487c701-kube-api-access-d5997\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807084 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-public-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807121 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-internal-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807183 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data-custom\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807215 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-combined-ca-bundle\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807289 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a54eb8d-8507-4f32-b2bf-936ea487c701-logs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807788 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a54eb8d-8507-4f32-b2bf-936ea487c701-logs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.807838 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.813278 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-public-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.818771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-internal-tls-certs\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.822333 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.823668 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-config-data-custom\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.825444 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a54eb8d-8507-4f32-b2bf-936ea487c701-combined-ca-bundle\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.830039 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5997\" (UniqueName: \"kubernetes.io/projected/7a54eb8d-8507-4f32-b2bf-936ea487c701-kube-api-access-d5997\") pod \"barbican-api-cbd859876-xpw6c\" (UID: \"7a54eb8d-8507-4f32-b2bf-936ea487c701\") " pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:19 crc kubenswrapper[4867]: I1212 08:16:19.884490 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:20 crc kubenswrapper[4867]: I1212 08:16:20.455907 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerStarted","Data":"779c779d65e05adf36356069eb5181278aee7aa5aa8451a70ffd22354874090b"} Dec 12 08:16:20 crc kubenswrapper[4867]: I1212 08:16:20.457700 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" event={"ID":"e41196ac-a7c9-47fa-9ca0-584a5fad129b","Type":"ContainerStarted","Data":"71dcc964c56b8a10c6fa4008b4e0e28fdeb93cc5cb05c9481a2fbd87e22dddc0"} Dec 12 08:16:20 crc kubenswrapper[4867]: I1212 08:16:20.457720 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" event={"ID":"e41196ac-a7c9-47fa-9ca0-584a5fad129b","Type":"ContainerStarted","Data":"309122c4d4db24a61e007948d44464684c228f1cbe87b10a3dac8cfd795741fa"} Dec 12 08:16:20 crc kubenswrapper[4867]: I1212 08:16:20.462454 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" event={"ID":"2ad1843a-0c00-40d2-95b5-4f164cc54053","Type":"ContainerStarted","Data":"98e9544e137bec1d9ca2ac845b629ca439450859970657d47d4930e75eb6c072"} Dec 12 08:16:20 crc kubenswrapper[4867]: I1212 08:16:20.537668 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cbd859876-xpw6c"] Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.473683 4867 generic.go:334] "Generic (PLEG): container finished" podID="1c579316-8df2-4a9f-a386-c2c97b101676" containerID="779c779d65e05adf36356069eb5181278aee7aa5aa8451a70ffd22354874090b" exitCode=0 Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.473751 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerDied","Data":"779c779d65e05adf36356069eb5181278aee7aa5aa8451a70ffd22354874090b"} Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.480548 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" event={"ID":"2ad1843a-0c00-40d2-95b5-4f164cc54053","Type":"ContainerStarted","Data":"41e59510dc62db0ab0b97fca62546b7d9ab88ca470a6be695775f4f38ebfdacd"} Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.488501 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbd859876-xpw6c" event={"ID":"7a54eb8d-8507-4f32-b2bf-936ea487c701","Type":"ContainerStarted","Data":"c9868cbe44ef190cbc742762582919c6f2eb0fb5e6d8ea3bf8750c1bf99916d6"} Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.489006 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.489026 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.489036 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbd859876-xpw6c" event={"ID":"7a54eb8d-8507-4f32-b2bf-936ea487c701","Type":"ContainerStarted","Data":"b1804d82e812ac321751b26a4a6dd54649a53e1b54086b6c742fa3a931c25016"} Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.489047 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cbd859876-xpw6c" event={"ID":"7a54eb8d-8507-4f32-b2bf-936ea487c701","Type":"ContainerStarted","Data":"f3b264ee51e79744cfba3b41c648f90ad5e54c2a04fd4c3397ab77a4b3427c22"} Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.520040 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5bff489498-h54sr" podStartSLOduration=2.88155222 podStartE2EDuration="5.520019816s" podCreationTimestamp="2025-12-12 08:16:16 +0000 UTC" firstStartedPulling="2025-12-12 08:16:17.371476527 +0000 UTC m=+5264.942857796" lastFinishedPulling="2025-12-12 08:16:20.009944123 +0000 UTC m=+5267.581325392" observedRunningTime="2025-12-12 08:16:21.518830857 +0000 UTC m=+5269.090212136" watchObservedRunningTime="2025-12-12 08:16:21.520019816 +0000 UTC m=+5269.091401085" Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.555867 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cbd859876-xpw6c" podStartSLOduration=2.555844237 podStartE2EDuration="2.555844237s" podCreationTimestamp="2025-12-12 08:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:16:21.5494827 +0000 UTC m=+5269.120863989" watchObservedRunningTime="2025-12-12 08:16:21.555844237 +0000 UTC m=+5269.127225506" Dec 12 08:16:21 crc kubenswrapper[4867]: I1212 08:16:21.574946 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84b67f6bf7-jqpkj" podStartSLOduration=3.031635197 podStartE2EDuration="5.574924365s" podCreationTimestamp="2025-12-12 08:16:16 +0000 UTC" firstStartedPulling="2025-12-12 08:16:17.452462656 +0000 UTC m=+5265.023843925" lastFinishedPulling="2025-12-12 08:16:19.995751824 +0000 UTC m=+5267.567133093" observedRunningTime="2025-12-12 08:16:21.569353189 +0000 UTC m=+5269.140734478" watchObservedRunningTime="2025-12-12 08:16:21.574924365 +0000 UTC m=+5269.146305634" Dec 12 08:16:22 crc kubenswrapper[4867]: I1212 08:16:22.498278 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerStarted","Data":"bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d"} Dec 12 08:16:22 crc kubenswrapper[4867]: I1212 08:16:22.522702 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdvfn" podStartSLOduration=2.86603165 podStartE2EDuration="6.522680049s" podCreationTimestamp="2025-12-12 08:16:16 +0000 UTC" firstStartedPulling="2025-12-12 08:16:18.402360034 +0000 UTC m=+5265.973741303" lastFinishedPulling="2025-12-12 08:16:22.059008433 +0000 UTC m=+5269.630389702" observedRunningTime="2025-12-12 08:16:22.516941069 +0000 UTC m=+5270.088322348" watchObservedRunningTime="2025-12-12 08:16:22.522680049 +0000 UTC m=+5270.094061318" Dec 12 08:16:26 crc kubenswrapper[4867]: I1212 08:16:26.968427 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.030891 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.031535 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="dnsmasq-dns" containerID="cri-o://eb9a1029e25d39c84f7a191d1a9ced6a9c9703852deb7449b67da9801e0a227a" gracePeriod=10 Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.343471 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.344014 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.387154 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.600354 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:27 crc kubenswrapper[4867]: I1212 08:16:27.653716 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.293394 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.18:5353: connect: connection refused" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.559219 4867 generic.go:334] "Generic (PLEG): container finished" podID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerID="eb9a1029e25d39c84f7a191d1a9ced6a9c9703852deb7449b67da9801e0a227a" exitCode=0 Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.560365 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" event={"ID":"036afb2d-c0e8-4f31-b0b8-e5548773af3c","Type":"ContainerDied","Data":"eb9a1029e25d39c84f7a191d1a9ced6a9c9703852deb7449b67da9801e0a227a"} Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.560399 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" event={"ID":"036afb2d-c0e8-4f31-b0b8-e5548773af3c","Type":"ContainerDied","Data":"e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591"} Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.560410 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e71eb342136f685789889fc4b97568d862f88f27670bb8e7886898147f55f591" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.618365 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.643131 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.674579 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.772172 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzljc\" (UniqueName: \"kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc\") pod \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.772314 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc\") pod \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.772435 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config\") pod \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.772515 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb\") pod \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.772542 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb\") pod \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\" (UID: \"036afb2d-c0e8-4f31-b0b8-e5548773af3c\") " Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.778265 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc" (OuterVolumeSpecName: "kube-api-access-kzljc") pod "036afb2d-c0e8-4f31-b0b8-e5548773af3c" (UID: "036afb2d-c0e8-4f31-b0b8-e5548773af3c"). InnerVolumeSpecName "kube-api-access-kzljc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.817636 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "036afb2d-c0e8-4f31-b0b8-e5548773af3c" (UID: "036afb2d-c0e8-4f31-b0b8-e5548773af3c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.834141 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config" (OuterVolumeSpecName: "config") pod "036afb2d-c0e8-4f31-b0b8-e5548773af3c" (UID: "036afb2d-c0e8-4f31-b0b8-e5548773af3c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.839576 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "036afb2d-c0e8-4f31-b0b8-e5548773af3c" (UID: "036afb2d-c0e8-4f31-b0b8-e5548773af3c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.849681 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "036afb2d-c0e8-4f31-b0b8-e5548773af3c" (UID: "036afb2d-c0e8-4f31-b0b8-e5548773af3c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.875667 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzljc\" (UniqueName: \"kubernetes.io/projected/036afb2d-c0e8-4f31-b0b8-e5548773af3c-kube-api-access-kzljc\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.875709 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.875721 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.875733 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.875743 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/036afb2d-c0e8-4f31-b0b8-e5548773af3c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.988705 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.989072 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.989122 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.989895 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:16:28 crc kubenswrapper[4867]: I1212 08:16:28.989968 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" gracePeriod=600 Dec 12 08:16:29 crc kubenswrapper[4867]: E1212 08:16:29.225316 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.588449 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" exitCode=0 Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.588506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570"} Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.588607 4867 scope.go:117] "RemoveContainer" containerID="59bd038a6c36ab860d64edc0db3240e73e8a95507655cbff9753c7840755dfbf" Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.588663 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5f75dd7-ptnt2" Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.589353 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:16:29 crc kubenswrapper[4867]: E1212 08:16:29.589665 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.589939 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdvfn" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="registry-server" containerID="cri-o://bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d" gracePeriod=2 Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.688706 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:16:29 crc kubenswrapper[4867]: I1212 08:16:29.709608 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76d5f75dd7-ptnt2"] Dec 12 08:16:29 crc kubenswrapper[4867]: E1212 08:16:29.797275 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c579316_8df2_4a9f_a386_c2c97b101676.slice/crio-conmon-bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.604631 4867 generic.go:334] "Generic (PLEG): container finished" podID="1c579316-8df2-4a9f-a386-c2c97b101676" containerID="bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d" exitCode=0 Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.604992 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerDied","Data":"bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d"} Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.605022 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdvfn" event={"ID":"1c579316-8df2-4a9f-a386-c2c97b101676","Type":"ContainerDied","Data":"a37b75bddca640ed3f304f24f4b0a327f9961f2f66ebd33a5986e41bafa828dd"} Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.605068 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a37b75bddca640ed3f304f24f4b0a327f9961f2f66ebd33a5986e41bafa828dd" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.659247 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.814116 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7q66\" (UniqueName: \"kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66\") pod \"1c579316-8df2-4a9f-a386-c2c97b101676\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.814334 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities\") pod \"1c579316-8df2-4a9f-a386-c2c97b101676\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.814419 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content\") pod \"1c579316-8df2-4a9f-a386-c2c97b101676\" (UID: \"1c579316-8df2-4a9f-a386-c2c97b101676\") " Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.815262 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities" (OuterVolumeSpecName: "utilities") pod "1c579316-8df2-4a9f-a386-c2c97b101676" (UID: "1c579316-8df2-4a9f-a386-c2c97b101676"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.820780 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66" (OuterVolumeSpecName: "kube-api-access-h7q66") pod "1c579316-8df2-4a9f-a386-c2c97b101676" (UID: "1c579316-8df2-4a9f-a386-c2c97b101676"). InnerVolumeSpecName "kube-api-access-h7q66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.848527 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" path="/var/lib/kubelet/pods/036afb2d-c0e8-4f31-b0b8-e5548773af3c/volumes" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.916540 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7q66\" (UniqueName: \"kubernetes.io/projected/1c579316-8df2-4a9f-a386-c2c97b101676-kube-api-access-h7q66\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:30 crc kubenswrapper[4867]: I1212 08:16:30.916581 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.461297 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.613471 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdvfn" Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.681756 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cbd859876-xpw6c" Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.792705 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.792932 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74f48cfc74-rblds" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api-log" containerID="cri-o://ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf" gracePeriod=30 Dec 12 08:16:31 crc kubenswrapper[4867]: I1212 08:16:31.793393 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74f48cfc74-rblds" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api" containerID="cri-o://8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253" gracePeriod=30 Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.086217 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c579316-8df2-4a9f-a386-c2c97b101676" (UID: "1c579316-8df2-4a9f-a386-c2c97b101676"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.130565 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c579316-8df2-4a9f-a386-c2c97b101676-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.144699 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.153133 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdvfn"] Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.628276 4867 generic.go:334] "Generic (PLEG): container finished" podID="b90aa2ea-887e-4671-b487-af95fea22d34" containerID="ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf" exitCode=143 Dec 12 08:16:33 crc kubenswrapper[4867]: I1212 08:16:33.628328 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerDied","Data":"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf"} Dec 12 08:16:34 crc kubenswrapper[4867]: I1212 08:16:34.848936 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" path="/var/lib/kubelet/pods/1c579316-8df2-4a9f-a386-c2c97b101676/volumes" Dec 12 08:16:34 crc kubenswrapper[4867]: I1212 08:16:34.943142 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74f48cfc74-rblds" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.34:9311/healthcheck\": read tcp 10.217.0.2:35446->10.217.1.34:9311: read: connection reset by peer" Dec 12 08:16:34 crc kubenswrapper[4867]: I1212 08:16:34.943142 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74f48cfc74-rblds" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.34:9311/healthcheck\": read tcp 10.217.0.2:35430->10.217.1.34:9311: read: connection reset by peer" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.356404 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.467682 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom\") pod \"b90aa2ea-887e-4671-b487-af95fea22d34\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.467905 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4wnj\" (UniqueName: \"kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj\") pod \"b90aa2ea-887e-4671-b487-af95fea22d34\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.467951 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data\") pod \"b90aa2ea-887e-4671-b487-af95fea22d34\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.468171 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs\") pod \"b90aa2ea-887e-4671-b487-af95fea22d34\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.468196 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle\") pod \"b90aa2ea-887e-4671-b487-af95fea22d34\" (UID: \"b90aa2ea-887e-4671-b487-af95fea22d34\") " Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.468705 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs" (OuterVolumeSpecName: "logs") pod "b90aa2ea-887e-4671-b487-af95fea22d34" (UID: "b90aa2ea-887e-4671-b487-af95fea22d34"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.473444 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj" (OuterVolumeSpecName: "kube-api-access-w4wnj") pod "b90aa2ea-887e-4671-b487-af95fea22d34" (UID: "b90aa2ea-887e-4671-b487-af95fea22d34"). InnerVolumeSpecName "kube-api-access-w4wnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.473950 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b90aa2ea-887e-4671-b487-af95fea22d34" (UID: "b90aa2ea-887e-4671-b487-af95fea22d34"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.491379 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b90aa2ea-887e-4671-b487-af95fea22d34" (UID: "b90aa2ea-887e-4671-b487-af95fea22d34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.512954 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data" (OuterVolumeSpecName: "config-data") pod "b90aa2ea-887e-4671-b487-af95fea22d34" (UID: "b90aa2ea-887e-4671-b487-af95fea22d34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.570267 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.570306 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4wnj\" (UniqueName: \"kubernetes.io/projected/b90aa2ea-887e-4671-b487-af95fea22d34-kube-api-access-w4wnj\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.570316 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.570326 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90aa2ea-887e-4671-b487-af95fea22d34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.570334 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b90aa2ea-887e-4671-b487-af95fea22d34-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.650825 4867 generic.go:334] "Generic (PLEG): container finished" podID="b90aa2ea-887e-4671-b487-af95fea22d34" containerID="8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253" exitCode=0 Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.650903 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74f48cfc74-rblds" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.650905 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerDied","Data":"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253"} Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.651067 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74f48cfc74-rblds" event={"ID":"b90aa2ea-887e-4671-b487-af95fea22d34","Type":"ContainerDied","Data":"bbe3e8aedf23d098d972a074b1a520afc812a44ee16dd09c50b64d23bec416c8"} Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.651105 4867 scope.go:117] "RemoveContainer" containerID="8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.677154 4867 scope.go:117] "RemoveContainer" containerID="ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.682922 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.690108 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-74f48cfc74-rblds"] Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.694383 4867 scope.go:117] "RemoveContainer" containerID="8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253" Dec 12 08:16:35 crc kubenswrapper[4867]: E1212 08:16:35.695332 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253\": container with ID starting with 8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253 not found: ID does not exist" containerID="8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.695377 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253"} err="failed to get container status \"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253\": rpc error: code = NotFound desc = could not find container \"8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253\": container with ID starting with 8e716f1b6f04dc5d24ca99871200a3203134afa9b870e6409c7631c031934253 not found: ID does not exist" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.695405 4867 scope.go:117] "RemoveContainer" containerID="ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf" Dec 12 08:16:35 crc kubenswrapper[4867]: E1212 08:16:35.695682 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf\": container with ID starting with ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf not found: ID does not exist" containerID="ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf" Dec 12 08:16:35 crc kubenswrapper[4867]: I1212 08:16:35.695723 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf"} err="failed to get container status \"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf\": rpc error: code = NotFound desc = could not find container \"ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf\": container with ID starting with ee637e0fbd8fd3f83e20d1d492a154966e44e54929c16dfccfb4dae2fdcf9fbf not found: ID does not exist" Dec 12 08:16:36 crc kubenswrapper[4867]: I1212 08:16:36.847792 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" path="/var/lib/kubelet/pods/b90aa2ea-887e-4671-b487-af95fea22d34/volumes" Dec 12 08:16:42 crc kubenswrapper[4867]: I1212 08:16:42.849167 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:16:42 crc kubenswrapper[4867]: E1212 08:16:42.850016 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.674946 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pjv5k"] Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675709 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="init" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675726 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="init" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675748 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="registry-server" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675756 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="registry-server" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675769 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="dnsmasq-dns" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675775 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="dnsmasq-dns" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675787 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="extract-utilities" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675793 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="extract-utilities" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675804 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api-log" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675810 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api-log" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675821 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="extract-content" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675826 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="extract-content" Dec 12 08:16:44 crc kubenswrapper[4867]: E1212 08:16:44.675846 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675853 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.675998 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="036afb2d-c0e8-4f31-b0b8-e5548773af3c" containerName="dnsmasq-dns" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.676008 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api-log" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.676028 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90aa2ea-887e-4671-b487-af95fea22d34" containerName="barbican-api" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.676039 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c579316-8df2-4a9f-a386-c2c97b101676" containerName="registry-server" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.676731 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.684021 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pjv5k"] Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.781869 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fe9d-account-create-update-hc4nz"] Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.782899 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.784614 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.793592 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fe9d-account-create-update-hc4nz"] Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.830268 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.830411 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbxg2\" (UniqueName: \"kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.932423 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.932497 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7qgb\" (UniqueName: \"kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.932527 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.932576 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbxg2\" (UniqueName: \"kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.933476 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.955580 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbxg2\" (UniqueName: \"kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2\") pod \"neutron-db-create-pjv5k\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:44 crc kubenswrapper[4867]: I1212 08:16:44.995266 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.034512 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7qgb\" (UniqueName: \"kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.034592 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.036109 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.056581 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7qgb\" (UniqueName: \"kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb\") pod \"neutron-fe9d-account-create-update-hc4nz\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.101811 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.506116 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pjv5k"] Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.524359 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fe9d-account-create-update-hc4nz"] Dec 12 08:16:45 crc kubenswrapper[4867]: W1212 08:16:45.543917 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3475cd4e_37de_44b2_8f2a_fb4f370b5a89.slice/crio-5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c WatchSource:0}: Error finding container 5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c: Status 404 returned error can't find the container with id 5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.725420 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pjv5k" event={"ID":"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03","Type":"ContainerStarted","Data":"ba8714818cf7f98e2c13048648d33131460928d92308394731ce5a7d3f4c8cf4"} Dec 12 08:16:45 crc kubenswrapper[4867]: I1212 08:16:45.727955 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fe9d-account-create-update-hc4nz" event={"ID":"3475cd4e-37de-44b2-8f2a-fb4f370b5a89","Type":"ContainerStarted","Data":"5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c"} Dec 12 08:16:46 crc kubenswrapper[4867]: I1212 08:16:46.737996 4867 generic.go:334] "Generic (PLEG): container finished" podID="3475cd4e-37de-44b2-8f2a-fb4f370b5a89" containerID="d0f25ff5dc033f4df4696b0cde0968038e55cb17617a44e5a10d86985545c6fc" exitCode=0 Dec 12 08:16:46 crc kubenswrapper[4867]: I1212 08:16:46.738060 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fe9d-account-create-update-hc4nz" event={"ID":"3475cd4e-37de-44b2-8f2a-fb4f370b5a89","Type":"ContainerDied","Data":"d0f25ff5dc033f4df4696b0cde0968038e55cb17617a44e5a10d86985545c6fc"} Dec 12 08:16:46 crc kubenswrapper[4867]: I1212 08:16:46.741175 4867 generic.go:334] "Generic (PLEG): container finished" podID="b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" containerID="0a8e5392a6aa0f51d7bce68e67536728957ae4533f9397c7491203c9c7bfd8d4" exitCode=0 Dec 12 08:16:46 crc kubenswrapper[4867]: I1212 08:16:46.741281 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pjv5k" event={"ID":"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03","Type":"ContainerDied","Data":"0a8e5392a6aa0f51d7bce68e67536728957ae4533f9397c7491203c9c7bfd8d4"} Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.158281 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.164911 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.293640 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts\") pod \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.293778 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7qgb\" (UniqueName: \"kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb\") pod \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.293836 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts\") pod \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\" (UID: \"3475cd4e-37de-44b2-8f2a-fb4f370b5a89\") " Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.294015 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbxg2\" (UniqueName: \"kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2\") pod \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\" (UID: \"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03\") " Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.294368 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" (UID: "b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.294479 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3475cd4e-37de-44b2-8f2a-fb4f370b5a89" (UID: "3475cd4e-37de-44b2-8f2a-fb4f370b5a89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.298966 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb" (OuterVolumeSpecName: "kube-api-access-z7qgb") pod "3475cd4e-37de-44b2-8f2a-fb4f370b5a89" (UID: "3475cd4e-37de-44b2-8f2a-fb4f370b5a89"). InnerVolumeSpecName "kube-api-access-z7qgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.299360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2" (OuterVolumeSpecName: "kube-api-access-xbxg2") pod "b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" (UID: "b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03"). InnerVolumeSpecName "kube-api-access-xbxg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.396528 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbxg2\" (UniqueName: \"kubernetes.io/projected/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-kube-api-access-xbxg2\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.396568 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.396580 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7qgb\" (UniqueName: \"kubernetes.io/projected/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-kube-api-access-z7qgb\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.396589 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3475cd4e-37de-44b2-8f2a-fb4f370b5a89-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.757539 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pjv5k" event={"ID":"b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03","Type":"ContainerDied","Data":"ba8714818cf7f98e2c13048648d33131460928d92308394731ce5a7d3f4c8cf4"} Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.757569 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pjv5k" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.757578 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba8714818cf7f98e2c13048648d33131460928d92308394731ce5a7d3f4c8cf4" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.759158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fe9d-account-create-update-hc4nz" event={"ID":"3475cd4e-37de-44b2-8f2a-fb4f370b5a89","Type":"ContainerDied","Data":"5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c"} Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.759182 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a3f8d63ec192b29c445a9b079f9696f2af1161856e6b85505c2e75168a5468c" Dec 12 08:16:48 crc kubenswrapper[4867]: I1212 08:16:48.759211 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fe9d-account-create-update-hc4nz" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.202334 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-j67rb"] Dec 12 08:16:50 crc kubenswrapper[4867]: E1212 08:16:50.203057 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" containerName="mariadb-database-create" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.203073 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" containerName="mariadb-database-create" Dec 12 08:16:50 crc kubenswrapper[4867]: E1212 08:16:50.203101 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3475cd4e-37de-44b2-8f2a-fb4f370b5a89" containerName="mariadb-account-create-update" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.203108 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3475cd4e-37de-44b2-8f2a-fb4f370b5a89" containerName="mariadb-account-create-update" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.203302 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" containerName="mariadb-database-create" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.203330 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3475cd4e-37de-44b2-8f2a-fb4f370b5a89" containerName="mariadb-account-create-update" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.203915 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.206906 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.207081 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.207273 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ksmhq" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.220417 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j67rb"] Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.328004 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.328116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.328216 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsdvf\" (UniqueName: \"kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.429915 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.430045 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.430131 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsdvf\" (UniqueName: \"kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.435414 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.441957 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.444873 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsdvf\" (UniqueName: \"kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf\") pod \"neutron-db-sync-j67rb\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.525717 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:50 crc kubenswrapper[4867]: I1212 08:16:50.957118 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-j67rb"] Dec 12 08:16:51 crc kubenswrapper[4867]: I1212 08:16:51.792649 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j67rb" event={"ID":"82193c94-00e1-47d0-ab01-992d27297d32","Type":"ContainerStarted","Data":"4e3809d01fc107522673b9e7d3500c16fe4c754545c20db485219df3a10ab2bc"} Dec 12 08:16:51 crc kubenswrapper[4867]: I1212 08:16:51.792964 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j67rb" event={"ID":"82193c94-00e1-47d0-ab01-992d27297d32","Type":"ContainerStarted","Data":"5925edcbaa7f7651dbf1ed8d76532e09313554f9b8ba4311e30721cec79eacf7"} Dec 12 08:16:51 crc kubenswrapper[4867]: I1212 08:16:51.810262 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-j67rb" podStartSLOduration=1.8102233349999999 podStartE2EDuration="1.810223335s" podCreationTimestamp="2025-12-12 08:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:16:51.806966675 +0000 UTC m=+5299.378347944" watchObservedRunningTime="2025-12-12 08:16:51.810223335 +0000 UTC m=+5299.381604614" Dec 12 08:16:53 crc kubenswrapper[4867]: I1212 08:16:53.837904 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:16:53 crc kubenswrapper[4867]: E1212 08:16:53.838532 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:16:55 crc kubenswrapper[4867]: I1212 08:16:55.825259 4867 generic.go:334] "Generic (PLEG): container finished" podID="82193c94-00e1-47d0-ab01-992d27297d32" containerID="4e3809d01fc107522673b9e7d3500c16fe4c754545c20db485219df3a10ab2bc" exitCode=0 Dec 12 08:16:55 crc kubenswrapper[4867]: I1212 08:16:55.825293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j67rb" event={"ID":"82193c94-00e1-47d0-ab01-992d27297d32","Type":"ContainerDied","Data":"4e3809d01fc107522673b9e7d3500c16fe4c754545c20db485219df3a10ab2bc"} Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.140839 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.244184 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle\") pod \"82193c94-00e1-47d0-ab01-992d27297d32\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.244330 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config\") pod \"82193c94-00e1-47d0-ab01-992d27297d32\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.244359 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsdvf\" (UniqueName: \"kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf\") pod \"82193c94-00e1-47d0-ab01-992d27297d32\" (UID: \"82193c94-00e1-47d0-ab01-992d27297d32\") " Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.250206 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf" (OuterVolumeSpecName: "kube-api-access-tsdvf") pod "82193c94-00e1-47d0-ab01-992d27297d32" (UID: "82193c94-00e1-47d0-ab01-992d27297d32"). InnerVolumeSpecName "kube-api-access-tsdvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.269934 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config" (OuterVolumeSpecName: "config") pod "82193c94-00e1-47d0-ab01-992d27297d32" (UID: "82193c94-00e1-47d0-ab01-992d27297d32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.273392 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82193c94-00e1-47d0-ab01-992d27297d32" (UID: "82193c94-00e1-47d0-ab01-992d27297d32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.346646 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.346698 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/82193c94-00e1-47d0-ab01-992d27297d32-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.346714 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsdvf\" (UniqueName: \"kubernetes.io/projected/82193c94-00e1-47d0-ab01-992d27297d32-kube-api-access-tsdvf\") on node \"crc\" DevicePath \"\"" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.842641 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-j67rb" event={"ID":"82193c94-00e1-47d0-ab01-992d27297d32","Type":"ContainerDied","Data":"5925edcbaa7f7651dbf1ed8d76532e09313554f9b8ba4311e30721cec79eacf7"} Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.842683 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5925edcbaa7f7651dbf1ed8d76532e09313554f9b8ba4311e30721cec79eacf7" Dec 12 08:16:57 crc kubenswrapper[4867]: I1212 08:16:57.842697 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-j67rb" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.065132 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:16:58 crc kubenswrapper[4867]: E1212 08:16:58.067724 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82193c94-00e1-47d0-ab01-992d27297d32" containerName="neutron-db-sync" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.067815 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="82193c94-00e1-47d0-ab01-992d27297d32" containerName="neutron-db-sync" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.068607 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="82193c94-00e1-47d0-ab01-992d27297d32" containerName="neutron-db-sync" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.070568 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.124486 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.170579 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.172639 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.172944 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.173007 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.173053 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.173129 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.173261 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqvkj\" (UniqueName: \"kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.177547 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.177767 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ksmhq" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.177994 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.178439 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.183595 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275754 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275813 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275890 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7kfd\" (UniqueName: \"kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275928 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275951 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.275982 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqvkj\" (UniqueName: \"kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.276020 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.276041 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.276062 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.276093 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.277029 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.277789 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.284970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.288378 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.303583 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqvkj\" (UniqueName: \"kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj\") pod \"dnsmasq-dns-5687f99fb5-tn724\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.377519 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.377632 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.377688 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7kfd\" (UniqueName: \"kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.377717 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.377738 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.382611 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.382729 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.383013 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.385775 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.397189 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7kfd\" (UniqueName: \"kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd\") pod \"neutron-d5d6d4d8d-8db6m\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.411135 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.493316 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:58 crc kubenswrapper[4867]: I1212 08:16:58.906253 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.135603 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:16:59 crc kubenswrapper[4867]: W1212 08:16:59.180911 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01d6156c_d90a_4336_9aa2_71c7df735fcf.slice/crio-53947109d5776ca667abe478fdc0c044b3d2ab8e154962167d6afb22a765710c WatchSource:0}: Error finding container 53947109d5776ca667abe478fdc0c044b3d2ab8e154962167d6afb22a765710c: Status 404 returned error can't find the container with id 53947109d5776ca667abe478fdc0c044b3d2ab8e154962167d6afb22a765710c Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.872485 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerStarted","Data":"375229a45cde82779e8e7fe312b013c79eaa97b982d0e4d00d4544cd09c1def3"} Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.872834 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerStarted","Data":"2a9398e4a50e8334104e102127468d65ab3d56ba4e39b4c687626f6f5f29ea76"} Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.872852 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.872861 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerStarted","Data":"53947109d5776ca667abe478fdc0c044b3d2ab8e154962167d6afb22a765710c"} Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.873587 4867 generic.go:334] "Generic (PLEG): container finished" podID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerID="0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a" exitCode=0 Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.873619 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" event={"ID":"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12","Type":"ContainerDied","Data":"0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a"} Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.873636 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" event={"ID":"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12","Type":"ContainerStarted","Data":"211dddac9438c07ea057234b515ea69dd11d016dcc9541d34ef257b15b727e32"} Dec 12 08:16:59 crc kubenswrapper[4867]: I1212 08:16:59.919325 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d5d6d4d8d-8db6m" podStartSLOduration=1.919304755 podStartE2EDuration="1.919304755s" podCreationTimestamp="2025-12-12 08:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:16:59.916814724 +0000 UTC m=+5307.488195993" watchObservedRunningTime="2025-12-12 08:16:59.919304755 +0000 UTC m=+5307.490686014" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.663600 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6c5858cbcf-825r6"] Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.665242 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.667570 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.667862 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.676388 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c5858cbcf-825r6"] Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824525 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-combined-ca-bundle\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824598 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55n5g\" (UniqueName: \"kubernetes.io/projected/f04c345a-28e4-412e-8d33-82d2f59586db-kube-api-access-55n5g\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824654 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-public-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824703 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-httpd-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824731 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-internal-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.824758 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-ovndb-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.895375 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" event={"ID":"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12","Type":"ContainerStarted","Data":"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39"} Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.895560 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.923588 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" podStartSLOduration=2.9235712080000003 podStartE2EDuration="2.923571208s" podCreationTimestamp="2025-12-12 08:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:17:00.915575761 +0000 UTC m=+5308.486957050" watchObservedRunningTime="2025-12-12 08:17:00.923571208 +0000 UTC m=+5308.494952477" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926440 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-httpd-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926500 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-internal-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926532 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-ovndb-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926597 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926688 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-combined-ca-bundle\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926812 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55n5g\" (UniqueName: \"kubernetes.io/projected/f04c345a-28e4-412e-8d33-82d2f59586db-kube-api-access-55n5g\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.926852 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-public-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.932664 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-combined-ca-bundle\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.933061 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-public-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.933196 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-ovndb-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.933372 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.933711 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-internal-tls-certs\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.942010 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f04c345a-28e4-412e-8d33-82d2f59586db-httpd-config\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.945482 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55n5g\" (UniqueName: \"kubernetes.io/projected/f04c345a-28e4-412e-8d33-82d2f59586db-kube-api-access-55n5g\") pod \"neutron-6c5858cbcf-825r6\" (UID: \"f04c345a-28e4-412e-8d33-82d2f59586db\") " pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:00 crc kubenswrapper[4867]: I1212 08:17:00.982635 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:01 crc kubenswrapper[4867]: I1212 08:17:01.519399 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6c5858cbcf-825r6"] Dec 12 08:17:01 crc kubenswrapper[4867]: I1212 08:17:01.908501 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5858cbcf-825r6" event={"ID":"f04c345a-28e4-412e-8d33-82d2f59586db","Type":"ContainerStarted","Data":"76e3e73e39de47b42e4ecf8dafc65c1e9bc6bd7140c2d0091037f0c068362ec4"} Dec 12 08:17:01 crc kubenswrapper[4867]: I1212 08:17:01.908878 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5858cbcf-825r6" event={"ID":"f04c345a-28e4-412e-8d33-82d2f59586db","Type":"ContainerStarted","Data":"2c775497f696cc0581b9470b2e28055b6be7840b5135f610f52f3fa93561035b"} Dec 12 08:17:02 crc kubenswrapper[4867]: I1212 08:17:02.920772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6c5858cbcf-825r6" event={"ID":"f04c345a-28e4-412e-8d33-82d2f59586db","Type":"ContainerStarted","Data":"6c61882ada94c1ebd5fdd768d24620fbbd57a9e3c26f084a01d6ee4f895abf55"} Dec 12 08:17:02 crc kubenswrapper[4867]: I1212 08:17:02.921415 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:02 crc kubenswrapper[4867]: I1212 08:17:02.949076 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6c5858cbcf-825r6" podStartSLOduration=2.949033638 podStartE2EDuration="2.949033638s" podCreationTimestamp="2025-12-12 08:17:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:17:02.940738905 +0000 UTC m=+5310.512120184" watchObservedRunningTime="2025-12-12 08:17:02.949033638 +0000 UTC m=+5310.520414907" Dec 12 08:17:05 crc kubenswrapper[4867]: I1212 08:17:05.838387 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:17:05 crc kubenswrapper[4867]: E1212 08:17:05.838967 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.412471 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.467874 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.468153 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="dnsmasq-dns" containerID="cri-o://dedf6ca4ad16f45eb6bc92f2324981d45457f8ecc3f587f476394ca47e89cb58" gracePeriod=10 Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.970965 4867 generic.go:334] "Generic (PLEG): container finished" podID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerID="dedf6ca4ad16f45eb6bc92f2324981d45457f8ecc3f587f476394ca47e89cb58" exitCode=0 Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.971043 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" event={"ID":"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2","Type":"ContainerDied","Data":"dedf6ca4ad16f45eb6bc92f2324981d45457f8ecc3f587f476394ca47e89cb58"} Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.971363 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" event={"ID":"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2","Type":"ContainerDied","Data":"77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b"} Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.971384 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77b658ae2e99f47aac916ea3b8f3d97769a84ad96705e1e68505b91af3637c7b" Dec 12 08:17:08 crc kubenswrapper[4867]: I1212 08:17:08.971638 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.061547 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb\") pod \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.061629 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc\") pod \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.061702 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config\") pod \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.061735 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd84j\" (UniqueName: \"kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j\") pod \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.061785 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb\") pod \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\" (UID: \"d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2\") " Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.067027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j" (OuterVolumeSpecName: "kube-api-access-sd84j") pod "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" (UID: "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2"). InnerVolumeSpecName "kube-api-access-sd84j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.106700 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" (UID: "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.112173 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" (UID: "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.116941 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config" (OuterVolumeSpecName: "config") pod "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" (UID: "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.122734 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" (UID: "d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.164194 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd84j\" (UniqueName: \"kubernetes.io/projected/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-kube-api-access-sd84j\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.164249 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.164266 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.164278 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.164289 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:09 crc kubenswrapper[4867]: I1212 08:17:09.979568 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc9d59f9-7jslg" Dec 12 08:17:10 crc kubenswrapper[4867]: I1212 08:17:10.031822 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:17:10 crc kubenswrapper[4867]: I1212 08:17:10.044887 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79dc9d59f9-7jslg"] Dec 12 08:17:10 crc kubenswrapper[4867]: I1212 08:17:10.850390 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" path="/var/lib/kubelet/pods/d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2/volumes" Dec 12 08:17:19 crc kubenswrapper[4867]: I1212 08:17:19.844487 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:17:19 crc kubenswrapper[4867]: E1212 08:17:19.846450 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:17:28 crc kubenswrapper[4867]: I1212 08:17:28.502155 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:17:30 crc kubenswrapper[4867]: I1212 08:17:30.838111 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:17:30 crc kubenswrapper[4867]: E1212 08:17:30.839115 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:17:31 crc kubenswrapper[4867]: I1212 08:17:31.001154 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6c5858cbcf-825r6" Dec 12 08:17:31 crc kubenswrapper[4867]: I1212 08:17:31.078862 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:17:31 crc kubenswrapper[4867]: I1212 08:17:31.079087 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d5d6d4d8d-8db6m" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-api" containerID="cri-o://2a9398e4a50e8334104e102127468d65ab3d56ba4e39b4c687626f6f5f29ea76" gracePeriod=30 Dec 12 08:17:31 crc kubenswrapper[4867]: I1212 08:17:31.079213 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-d5d6d4d8d-8db6m" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-httpd" containerID="cri-o://375229a45cde82779e8e7fe312b013c79eaa97b982d0e4d00d4544cd09c1def3" gracePeriod=30 Dec 12 08:17:32 crc kubenswrapper[4867]: I1212 08:17:32.166381 4867 generic.go:334] "Generic (PLEG): container finished" podID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerID="375229a45cde82779e8e7fe312b013c79eaa97b982d0e4d00d4544cd09c1def3" exitCode=0 Dec 12 08:17:32 crc kubenswrapper[4867]: I1212 08:17:32.166519 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerDied","Data":"375229a45cde82779e8e7fe312b013c79eaa97b982d0e4d00d4544cd09c1def3"} Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.220255 4867 generic.go:334] "Generic (PLEG): container finished" podID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerID="2a9398e4a50e8334104e102127468d65ab3d56ba4e39b4c687626f6f5f29ea76" exitCode=0 Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.220297 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerDied","Data":"2a9398e4a50e8334104e102127468d65ab3d56ba4e39b4c687626f6f5f29ea76"} Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.743313 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.786349 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7kfd\" (UniqueName: \"kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd\") pod \"01d6156c-d90a-4336-9aa2-71c7df735fcf\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.786458 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config\") pod \"01d6156c-d90a-4336-9aa2-71c7df735fcf\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.786516 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs\") pod \"01d6156c-d90a-4336-9aa2-71c7df735fcf\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.786594 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle\") pod \"01d6156c-d90a-4336-9aa2-71c7df735fcf\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.786687 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config\") pod \"01d6156c-d90a-4336-9aa2-71c7df735fcf\" (UID: \"01d6156c-d90a-4336-9aa2-71c7df735fcf\") " Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.795363 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "01d6156c-d90a-4336-9aa2-71c7df735fcf" (UID: "01d6156c-d90a-4336-9aa2-71c7df735fcf"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.802294 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd" (OuterVolumeSpecName: "kube-api-access-t7kfd") pod "01d6156c-d90a-4336-9aa2-71c7df735fcf" (UID: "01d6156c-d90a-4336-9aa2-71c7df735fcf"). InnerVolumeSpecName "kube-api-access-t7kfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.837393 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config" (OuterVolumeSpecName: "config") pod "01d6156c-d90a-4336-9aa2-71c7df735fcf" (UID: "01d6156c-d90a-4336-9aa2-71c7df735fcf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.844561 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01d6156c-d90a-4336-9aa2-71c7df735fcf" (UID: "01d6156c-d90a-4336-9aa2-71c7df735fcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.858486 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "01d6156c-d90a-4336-9aa2-71c7df735fcf" (UID: "01d6156c-d90a-4336-9aa2-71c7df735fcf"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.888586 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.888629 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7kfd\" (UniqueName: \"kubernetes.io/projected/01d6156c-d90a-4336-9aa2-71c7df735fcf-kube-api-access-t7kfd\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.888645 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.888657 4867 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:37 crc kubenswrapper[4867]: I1212 08:17:37.888671 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d6156c-d90a-4336-9aa2-71c7df735fcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.231449 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d5d6d4d8d-8db6m" event={"ID":"01d6156c-d90a-4336-9aa2-71c7df735fcf","Type":"ContainerDied","Data":"53947109d5776ca667abe478fdc0c044b3d2ab8e154962167d6afb22a765710c"} Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.231515 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d5d6d4d8d-8db6m" Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.231522 4867 scope.go:117] "RemoveContainer" containerID="375229a45cde82779e8e7fe312b013c79eaa97b982d0e4d00d4544cd09c1def3" Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.259848 4867 scope.go:117] "RemoveContainer" containerID="2a9398e4a50e8334104e102127468d65ab3d56ba4e39b4c687626f6f5f29ea76" Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.285814 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.291553 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d5d6d4d8d-8db6m"] Dec 12 08:17:38 crc kubenswrapper[4867]: I1212 08:17:38.850415 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" path="/var/lib/kubelet/pods/01d6156c-d90a-4336-9aa2-71c7df735fcf/volumes" Dec 12 08:17:45 crc kubenswrapper[4867]: I1212 08:17:45.838013 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:17:45 crc kubenswrapper[4867]: E1212 08:17:45.838757 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.248117 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-k2p9r"] Dec 12 08:17:50 crc kubenswrapper[4867]: E1212 08:17:50.258452 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-api" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258485 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-api" Dec 12 08:17:50 crc kubenswrapper[4867]: E1212 08:17:50.258526 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="init" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258532 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="init" Dec 12 08:17:50 crc kubenswrapper[4867]: E1212 08:17:50.258579 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="dnsmasq-dns" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258589 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="dnsmasq-dns" Dec 12 08:17:50 crc kubenswrapper[4867]: E1212 08:17:50.258610 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-httpd" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258618 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-httpd" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258925 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-api" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258955 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4345a89-22d4-4cb2-a6d2-dcb4efe1e1d2" containerName="dnsmasq-dns" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.258968 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="01d6156c-d90a-4336-9aa2-71c7df735fcf" containerName="neutron-httpd" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.259598 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.261608 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.262681 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.262786 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.263065 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-54lqq" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.265161 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.266396 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-k2p9r"] Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321459 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321513 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321560 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321609 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321633 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321670 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvb5c\" (UniqueName: \"kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.321696 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423257 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423340 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423378 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423425 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423493 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423524 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.423590 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvb5c\" (UniqueName: \"kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.424383 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.425086 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.425862 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.434149 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.436651 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.436807 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.455991 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvb5c\" (UniqueName: \"kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c\") pod \"swift-ring-rebalance-k2p9r\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.474061 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.476532 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.495155 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.525598 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.525688 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jr5\" (UniqueName: \"kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.525821 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.525856 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.525899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.579458 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.626970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.627018 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.627047 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.627118 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.627156 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6jr5\" (UniqueName: \"kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.628247 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.628789 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.629750 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.629936 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.649395 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6jr5\" (UniqueName: \"kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5\") pod \"dnsmasq-dns-5cb578ccbc-9pmn9\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:50 crc kubenswrapper[4867]: I1212 08:17:50.854126 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:51 crc kubenswrapper[4867]: I1212 08:17:51.131621 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-k2p9r"] Dec 12 08:17:51 crc kubenswrapper[4867]: I1212 08:17:51.360436 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k2p9r" event={"ID":"b54aaf28-10a9-4b58-b752-278929bf0bb8","Type":"ContainerStarted","Data":"60cccb3e129382020519f13c2a9d19109e3fa645dca841403dd013d67a96be88"} Dec 12 08:17:51 crc kubenswrapper[4867]: I1212 08:17:51.410974 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:17:52 crc kubenswrapper[4867]: I1212 08:17:52.373445 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerID="b241a464b00969263db9561d612d8fefa8e4c98024d093e712a197461c20da2d" exitCode=0 Dec 12 08:17:52 crc kubenswrapper[4867]: I1212 08:17:52.373519 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" event={"ID":"4a869be7-fa0e-4b99-8497-1e48c48acbff","Type":"ContainerDied","Data":"b241a464b00969263db9561d612d8fefa8e4c98024d093e712a197461c20da2d"} Dec 12 08:17:52 crc kubenswrapper[4867]: I1212 08:17:52.373975 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" event={"ID":"4a869be7-fa0e-4b99-8497-1e48c48acbff","Type":"ContainerStarted","Data":"59b0dd7f055a999cf6bf94a297ea07794fa6a9c3e4f28dda673e6b172e7e7621"} Dec 12 08:17:53 crc kubenswrapper[4867]: I1212 08:17:53.386616 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" event={"ID":"4a869be7-fa0e-4b99-8497-1e48c48acbff","Type":"ContainerStarted","Data":"d38185e32cb85a09dd8e447653f266d816a91a8996acc81ea5e31e825dba6291"} Dec 12 08:17:53 crc kubenswrapper[4867]: I1212 08:17:53.387205 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:17:53 crc kubenswrapper[4867]: I1212 08:17:53.406333 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" podStartSLOduration=3.406309813 podStartE2EDuration="3.406309813s" podCreationTimestamp="2025-12-12 08:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:17:53.405096633 +0000 UTC m=+5360.976477912" watchObservedRunningTime="2025-12-12 08:17:53.406309813 +0000 UTC m=+5360.977691082" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.242801 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6db4df84b6-rncss"] Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.244669 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.247685 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.247841 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.248104 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.252916 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6db4df84b6-rncss"] Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300573 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-config-data\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300688 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-combined-ca-bundle\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300812 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-run-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300886 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-log-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300938 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-etc-swift\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.300978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-internal-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.301037 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-public-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.301105 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4qv\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-kube-api-access-nq4qv\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402364 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-combined-ca-bundle\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402454 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-run-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402497 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-log-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402523 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-etc-swift\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402542 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-internal-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402587 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-public-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402625 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4qv\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-kube-api-access-nq4qv\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.402656 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-config-data\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.405097 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-run-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.405469 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3066d921-d258-4f35-b6fe-2dc9203299e5-log-httpd\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.408278 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-config-data\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.408290 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-internal-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.411785 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-public-tls-certs\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.412103 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3066d921-d258-4f35-b6fe-2dc9203299e5-combined-ca-bundle\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.420287 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-etc-swift\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.432449 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4qv\" (UniqueName: \"kubernetes.io/projected/3066d921-d258-4f35-b6fe-2dc9203299e5-kube-api-access-nq4qv\") pod \"swift-proxy-6db4df84b6-rncss\" (UID: \"3066d921-d258-4f35-b6fe-2dc9203299e5\") " pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:54 crc kubenswrapper[4867]: I1212 08:17:54.567779 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:56 crc kubenswrapper[4867]: I1212 08:17:56.384630 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6db4df84b6-rncss"] Dec 12 08:17:56 crc kubenswrapper[4867]: I1212 08:17:56.413577 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k2p9r" event={"ID":"b54aaf28-10a9-4b58-b752-278929bf0bb8","Type":"ContainerStarted","Data":"5da6021a44722ec3594fdba9f509ec20ab3549d881d157389c3c2335e9d29aef"} Dec 12 08:17:56 crc kubenswrapper[4867]: I1212 08:17:56.415653 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6db4df84b6-rncss" event={"ID":"3066d921-d258-4f35-b6fe-2dc9203299e5","Type":"ContainerStarted","Data":"4d5e55f6788decbc213b52cd209cc9e79e0f782e0ba7d6cc4bdf76963481e514"} Dec 12 08:17:56 crc kubenswrapper[4867]: I1212 08:17:56.434215 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-k2p9r" podStartSLOduration=1.757904802 podStartE2EDuration="6.43419243s" podCreationTimestamp="2025-12-12 08:17:50 +0000 UTC" firstStartedPulling="2025-12-12 08:17:51.121031852 +0000 UTC m=+5358.692413121" lastFinishedPulling="2025-12-12 08:17:55.79731948 +0000 UTC m=+5363.368700749" observedRunningTime="2025-12-12 08:17:56.428536071 +0000 UTC m=+5363.999917340" watchObservedRunningTime="2025-12-12 08:17:56.43419243 +0000 UTC m=+5364.005573699" Dec 12 08:17:57 crc kubenswrapper[4867]: I1212 08:17:57.427071 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6db4df84b6-rncss" event={"ID":"3066d921-d258-4f35-b6fe-2dc9203299e5","Type":"ContainerStarted","Data":"50f781eb35c717989099e54ffc39ad997385e2e61b8fce428afb8aeefc0df9e1"} Dec 12 08:17:57 crc kubenswrapper[4867]: I1212 08:17:57.427644 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6db4df84b6-rncss" event={"ID":"3066d921-d258-4f35-b6fe-2dc9203299e5","Type":"ContainerStarted","Data":"052112ced28a423e9345fe8590416b4c8957a9aae8fec1c6927e3575a2221daf"} Dec 12 08:17:57 crc kubenswrapper[4867]: I1212 08:17:57.465960 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6db4df84b6-rncss" podStartSLOduration=3.465942438 podStartE2EDuration="3.465942438s" podCreationTimestamp="2025-12-12 08:17:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:17:57.450465618 +0000 UTC m=+5365.021846907" watchObservedRunningTime="2025-12-12 08:17:57.465942438 +0000 UTC m=+5365.037323707" Dec 12 08:17:58 crc kubenswrapper[4867]: I1212 08:17:58.434430 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:58 crc kubenswrapper[4867]: I1212 08:17:58.434480 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:17:58 crc kubenswrapper[4867]: I1212 08:17:58.838593 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:17:58 crc kubenswrapper[4867]: E1212 08:17:58.838999 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:18:00 crc kubenswrapper[4867]: I1212 08:18:00.453722 4867 generic.go:334] "Generic (PLEG): container finished" podID="b54aaf28-10a9-4b58-b752-278929bf0bb8" containerID="5da6021a44722ec3594fdba9f509ec20ab3549d881d157389c3c2335e9d29aef" exitCode=0 Dec 12 08:18:00 crc kubenswrapper[4867]: I1212 08:18:00.453806 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k2p9r" event={"ID":"b54aaf28-10a9-4b58-b752-278929bf0bb8","Type":"ContainerDied","Data":"5da6021a44722ec3594fdba9f509ec20ab3549d881d157389c3c2335e9d29aef"} Dec 12 08:18:00 crc kubenswrapper[4867]: I1212 08:18:00.856019 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:18:00 crc kubenswrapper[4867]: I1212 08:18:00.928963 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:18:00 crc kubenswrapper[4867]: I1212 08:18:00.929326 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="dnsmasq-dns" containerID="cri-o://5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39" gracePeriod=10 Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.454269 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.462739 4867 generic.go:334] "Generic (PLEG): container finished" podID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerID="5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39" exitCode=0 Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.462814 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.462812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" event={"ID":"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12","Type":"ContainerDied","Data":"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39"} Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.462870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5687f99fb5-tn724" event={"ID":"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12","Type":"ContainerDied","Data":"211dddac9438c07ea057234b515ea69dd11d016dcc9541d34ef257b15b727e32"} Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.462893 4867 scope.go:117] "RemoveContainer" containerID="5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.506626 4867 scope.go:117] "RemoveContainer" containerID="0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.561220 4867 scope.go:117] "RemoveContainer" containerID="5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39" Dec 12 08:18:01 crc kubenswrapper[4867]: E1212 08:18:01.563792 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39\": container with ID starting with 5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39 not found: ID does not exist" containerID="5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.563846 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39"} err="failed to get container status \"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39\": rpc error: code = NotFound desc = could not find container \"5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39\": container with ID starting with 5c363141597dd00c6bbb473f45b1399e87aa206d2149971fe977947ba6fe4d39 not found: ID does not exist" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.563878 4867 scope.go:117] "RemoveContainer" containerID="0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a" Dec 12 08:18:01 crc kubenswrapper[4867]: E1212 08:18:01.565023 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a\": container with ID starting with 0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a not found: ID does not exist" containerID="0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.565079 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a"} err="failed to get container status \"0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a\": rpc error: code = NotFound desc = could not find container \"0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a\": container with ID starting with 0def274d707d216b2f1ae057eef16122af42fe2c27e5df2b9d8a9715c90f219a not found: ID does not exist" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.628264 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb\") pod \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.628428 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb\") pod \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.628488 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config\") pod \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.628555 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc\") pod \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.628581 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqvkj\" (UniqueName: \"kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj\") pod \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\" (UID: \"82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.681193 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj" (OuterVolumeSpecName: "kube-api-access-wqvkj") pod "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" (UID: "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12"). InnerVolumeSpecName "kube-api-access-wqvkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.709421 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" (UID: "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.718385 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" (UID: "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.722350 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config" (OuterVolumeSpecName: "config") pod "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" (UID: "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.732811 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.732853 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqvkj\" (UniqueName: \"kubernetes.io/projected/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-kube-api-access-wqvkj\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.732865 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.732878 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.742916 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" (UID: "82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.797462 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.812624 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.826936 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5687f99fb5-tn724"] Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.833703 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:01 crc kubenswrapper[4867]: E1212 08:18:01.924210 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82a9550c_a3c8_4ea9_a83b_b0d03bb0cd12.slice\": RecentStats: unable to find data in memory cache]" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.934815 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936280 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936415 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936571 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvb5c\" (UniqueName: \"kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936745 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936862 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.936985 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle\") pod \"b54aaf28-10a9-4b58-b752-278929bf0bb8\" (UID: \"b54aaf28-10a9-4b58-b752-278929bf0bb8\") " Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.935897 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.939304 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.941296 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c" (OuterVolumeSpecName: "kube-api-access-tvb5c") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "kube-api-access-tvb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.944999 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.960598 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts" (OuterVolumeSpecName: "scripts") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.962926 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:18:01 crc kubenswrapper[4867]: I1212 08:18:01.966849 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b54aaf28-10a9-4b58-b752-278929bf0bb8" (UID: "b54aaf28-10a9-4b58-b752-278929bf0bb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041571 4867 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041612 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041625 4867 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041636 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b54aaf28-10a9-4b58-b752-278929bf0bb8-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041646 4867 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b54aaf28-10a9-4b58-b752-278929bf0bb8-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041656 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvb5c\" (UniqueName: \"kubernetes.io/projected/b54aaf28-10a9-4b58-b752-278929bf0bb8-kube-api-access-tvb5c\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.041668 4867 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b54aaf28-10a9-4b58-b752-278929bf0bb8-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.475024 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-k2p9r" event={"ID":"b54aaf28-10a9-4b58-b752-278929bf0bb8","Type":"ContainerDied","Data":"60cccb3e129382020519f13c2a9d19109e3fa645dca841403dd013d67a96be88"} Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.476128 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60cccb3e129382020519f13c2a9d19109e3fa645dca841403dd013d67a96be88" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.475240 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-k2p9r" Dec 12 08:18:02 crc kubenswrapper[4867]: I1212 08:18:02.847837 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" path="/var/lib/kubelet/pods/82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12/volumes" Dec 12 08:18:04 crc kubenswrapper[4867]: I1212 08:18:04.576757 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:18:04 crc kubenswrapper[4867]: I1212 08:18:04.577631 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6db4df84b6-rncss" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.643785 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-27d8l"] Dec 12 08:18:10 crc kubenswrapper[4867]: E1212 08:18:10.645654 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="dnsmasq-dns" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.645676 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="dnsmasq-dns" Dec 12 08:18:10 crc kubenswrapper[4867]: E1212 08:18:10.645701 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b54aaf28-10a9-4b58-b752-278929bf0bb8" containerName="swift-ring-rebalance" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.645707 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b54aaf28-10a9-4b58-b752-278929bf0bb8" containerName="swift-ring-rebalance" Dec 12 08:18:10 crc kubenswrapper[4867]: E1212 08:18:10.645727 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="init" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.645732 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="init" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.646014 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a9550c-a3c8-4ea9-a83b-b0d03bb0cd12" containerName="dnsmasq-dns" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.646029 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b54aaf28-10a9-4b58-b752-278929bf0bb8" containerName="swift-ring-rebalance" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.646690 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.654106 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-27d8l"] Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.749060 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1932-account-create-update-kgphl"] Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.750134 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.753841 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.759138 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1932-account-create-update-kgphl"] Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.789271 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.789699 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7x72\" (UniqueName: \"kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.891359 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.891441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh2fz\" (UniqueName: \"kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.891579 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7x72\" (UniqueName: \"kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.891906 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.892451 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.921374 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7x72\" (UniqueName: \"kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72\") pod \"cinder-db-create-27d8l\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.983569 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.993118 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh2fz\" (UniqueName: \"kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.993326 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:10 crc kubenswrapper[4867]: I1212 08:18:10.993996 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.021082 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh2fz\" (UniqueName: \"kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz\") pod \"cinder-1932-account-create-update-kgphl\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.080939 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.441567 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-27d8l"] Dec 12 08:18:11 crc kubenswrapper[4867]: W1212 08:18:11.449998 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded7e6e15_71a3_41f0_a41e_4eae2d0fec66.slice/crio-49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab WatchSource:0}: Error finding container 49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab: Status 404 returned error can't find the container with id 49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.574822 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1932-account-create-update-kgphl"] Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.574876 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-27d8l" event={"ID":"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66","Type":"ContainerStarted","Data":"49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab"} Dec 12 08:18:11 crc kubenswrapper[4867]: W1212 08:18:11.575359 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e3e4886_12ff_4e96_b20e_7c13c2d54ce1.slice/crio-981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8 WatchSource:0}: Error finding container 981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8: Status 404 returned error can't find the container with id 981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8 Dec 12 08:18:11 crc kubenswrapper[4867]: I1212 08:18:11.838398 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:18:11 crc kubenswrapper[4867]: E1212 08:18:11.838981 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:18:12 crc kubenswrapper[4867]: I1212 08:18:12.584703 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1932-account-create-update-kgphl" event={"ID":"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1","Type":"ContainerStarted","Data":"981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8"} Dec 12 08:18:13 crc kubenswrapper[4867]: I1212 08:18:13.594404 4867 generic.go:334] "Generic (PLEG): container finished" podID="0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" containerID="4d2fb0d826d4a9896c79df299757c9549de17bdd402f90896ffc21c6a125bcc2" exitCode=0 Dec 12 08:18:13 crc kubenswrapper[4867]: I1212 08:18:13.594776 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1932-account-create-update-kgphl" event={"ID":"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1","Type":"ContainerDied","Data":"4d2fb0d826d4a9896c79df299757c9549de17bdd402f90896ffc21c6a125bcc2"} Dec 12 08:18:13 crc kubenswrapper[4867]: I1212 08:18:13.596668 4867 generic.go:334] "Generic (PLEG): container finished" podID="ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" containerID="791d9cdf6c3dfa77fa27cfa1d58814de36255d73326ecaafc877cf3bba8c3e96" exitCode=0 Dec 12 08:18:13 crc kubenswrapper[4867]: I1212 08:18:13.596702 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-27d8l" event={"ID":"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66","Type":"ContainerDied","Data":"791d9cdf6c3dfa77fa27cfa1d58814de36255d73326ecaafc877cf3bba8c3e96"} Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.005941 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.021965 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.170093 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh2fz\" (UniqueName: \"kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz\") pod \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.170176 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts\") pod \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\" (UID: \"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1\") " Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.170351 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7x72\" (UniqueName: \"kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72\") pod \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.170415 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts\") pod \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\" (UID: \"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66\") " Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.170854 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" (UID: "0e3e4886-12ff-4e96-b20e-7c13c2d54ce1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.171249 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" (UID: "ed7e6e15-71a3-41f0-a41e-4eae2d0fec66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.172060 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.172091 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.186538 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72" (OuterVolumeSpecName: "kube-api-access-n7x72") pod "ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" (UID: "ed7e6e15-71a3-41f0-a41e-4eae2d0fec66"). InnerVolumeSpecName "kube-api-access-n7x72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.191357 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz" (OuterVolumeSpecName: "kube-api-access-nh2fz") pod "0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" (UID: "0e3e4886-12ff-4e96-b20e-7c13c2d54ce1"). InnerVolumeSpecName "kube-api-access-nh2fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.274517 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh2fz\" (UniqueName: \"kubernetes.io/projected/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1-kube-api-access-nh2fz\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.274570 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7x72\" (UniqueName: \"kubernetes.io/projected/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66-kube-api-access-n7x72\") on node \"crc\" DevicePath \"\"" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.616573 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1932-account-create-update-kgphl" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.616561 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1932-account-create-update-kgphl" event={"ID":"0e3e4886-12ff-4e96-b20e-7c13c2d54ce1","Type":"ContainerDied","Data":"981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8"} Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.616781 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="981f89a78182709ffb0387e3d477a0090b8e4cc8049f14a65bec88af844a72d8" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.618520 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-27d8l" event={"ID":"ed7e6e15-71a3-41f0-a41e-4eae2d0fec66","Type":"ContainerDied","Data":"49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab"} Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.618558 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49983baa4c8e52663b2a5954df71f38244b7b974c49ce39394892a0921377cab" Dec 12 08:18:15 crc kubenswrapper[4867]: I1212 08:18:15.618591 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-27d8l" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.007206 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vpzt4"] Dec 12 08:18:21 crc kubenswrapper[4867]: E1212 08:18:21.009073 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" containerName="mariadb-account-create-update" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.009150 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" containerName="mariadb-account-create-update" Dec 12 08:18:21 crc kubenswrapper[4867]: E1212 08:18:21.009290 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" containerName="mariadb-database-create" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.009368 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" containerName="mariadb-database-create" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.009575 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" containerName="mariadb-account-create-update" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.009739 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" containerName="mariadb-database-create" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.010437 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.014094 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.014285 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.014549 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nb7t6" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.020448 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vpzt4"] Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070142 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070204 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070252 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9426\" (UniqueName: \"kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.070492 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172354 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172441 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172524 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172559 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172585 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9426\" (UniqueName: \"kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172628 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.172777 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.179024 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.179070 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.179464 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.179793 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.196907 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9426\" (UniqueName: \"kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426\") pod \"cinder-db-sync-vpzt4\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.389552 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:18:21 crc kubenswrapper[4867]: I1212 08:18:21.825351 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vpzt4"] Dec 12 08:18:22 crc kubenswrapper[4867]: I1212 08:18:22.677360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vpzt4" event={"ID":"17f2c606-4920-493b-a23d-b9b5209f3636","Type":"ContainerStarted","Data":"9f26411807f6430e25baa118e3e67ca723540e369aca7a4aaa1f69cfd151b852"} Dec 12 08:18:22 crc kubenswrapper[4867]: I1212 08:18:22.843372 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:18:22 crc kubenswrapper[4867]: E1212 08:18:22.843967 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:18:36 crc kubenswrapper[4867]: I1212 08:18:36.838686 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:18:36 crc kubenswrapper[4867]: E1212 08:18:36.839496 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:18:42 crc kubenswrapper[4867]: I1212 08:18:42.887072 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cbd859876-xpw6c" podUID="7a54eb8d-8507-4f32-b2bf-936ea487c701" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.1.36:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 12 08:18:42 crc kubenswrapper[4867]: I1212 08:18:42.888432 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cbd859876-xpw6c" podUID="7a54eb8d-8507-4f32-b2bf-936ea487c701" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.1.36:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 12 08:18:46 crc kubenswrapper[4867]: I1212 08:18:46.122850 4867 scope.go:117] "RemoveContainer" containerID="113dd68440afbde4000b8c6e3b5f4297d39cce26c25685759fd310153d2d57a4" Dec 12 08:18:51 crc kubenswrapper[4867]: I1212 08:18:51.838694 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:18:51 crc kubenswrapper[4867]: E1212 08:18:51.839709 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:18:53 crc kubenswrapper[4867]: E1212 08:18:53.590892 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:0a4adc83d85cf6902627a7ff9891be56" Dec 12 08:18:53 crc kubenswrapper[4867]: E1212 08:18:53.591432 4867 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:0a4adc83d85cf6902627a7ff9891be56" Dec 12 08:18:53 crc kubenswrapper[4867]: E1212 08:18:53.591633 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:0a4adc83d85cf6902627a7ff9891be56,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x9426,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vpzt4_openstack(17f2c606-4920-493b-a23d-b9b5209f3636): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 08:18:53 crc kubenswrapper[4867]: E1212 08:18:53.592816 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vpzt4" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" Dec 12 08:18:54 crc kubenswrapper[4867]: E1212 08:18:54.004088 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:0a4adc83d85cf6902627a7ff9891be56\\\"\"" pod="openstack/cinder-db-sync-vpzt4" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" Dec 12 08:19:02 crc kubenswrapper[4867]: I1212 08:19:02.846818 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:19:02 crc kubenswrapper[4867]: E1212 08:19:02.848337 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:19:06 crc kubenswrapper[4867]: I1212 08:19:06.096566 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vpzt4" event={"ID":"17f2c606-4920-493b-a23d-b9b5209f3636","Type":"ContainerStarted","Data":"18fc7f62366a413ad09766fb175f37064fcd6f4998d7d3dc26af29e6825a741d"} Dec 12 08:19:06 crc kubenswrapper[4867]: I1212 08:19:06.124435 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vpzt4" podStartSLOduration=2.925739922 podStartE2EDuration="46.124411612s" podCreationTimestamp="2025-12-12 08:18:20 +0000 UTC" firstStartedPulling="2025-12-12 08:18:21.842601772 +0000 UTC m=+5389.413983041" lastFinishedPulling="2025-12-12 08:19:05.041273462 +0000 UTC m=+5432.612654731" observedRunningTime="2025-12-12 08:19:06.114644122 +0000 UTC m=+5433.686025421" watchObservedRunningTime="2025-12-12 08:19:06.124411612 +0000 UTC m=+5433.695792891" Dec 12 08:19:08 crc kubenswrapper[4867]: I1212 08:19:08.117409 4867 generic.go:334] "Generic (PLEG): container finished" podID="17f2c606-4920-493b-a23d-b9b5209f3636" containerID="18fc7f62366a413ad09766fb175f37064fcd6f4998d7d3dc26af29e6825a741d" exitCode=0 Dec 12 08:19:08 crc kubenswrapper[4867]: I1212 08:19:08.117453 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vpzt4" event={"ID":"17f2c606-4920-493b-a23d-b9b5209f3636","Type":"ContainerDied","Data":"18fc7f62366a413ad09766fb175f37064fcd6f4998d7d3dc26af29e6825a741d"} Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.420209 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600691 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9426\" (UniqueName: \"kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600775 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600840 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600879 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600911 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.600959 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data\") pod \"17f2c606-4920-493b-a23d-b9b5209f3636\" (UID: \"17f2c606-4920-493b-a23d-b9b5209f3636\") " Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.601419 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.609359 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.610104 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts" (OuterVolumeSpecName: "scripts") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.612273 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426" (OuterVolumeSpecName: "kube-api-access-x9426") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "kube-api-access-x9426". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.633878 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.657575 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data" (OuterVolumeSpecName: "config-data") pod "17f2c606-4920-493b-a23d-b9b5209f3636" (UID: "17f2c606-4920-493b-a23d-b9b5209f3636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703524 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9426\" (UniqueName: \"kubernetes.io/projected/17f2c606-4920-493b-a23d-b9b5209f3636-kube-api-access-x9426\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703574 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703587 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703598 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/17f2c606-4920-493b-a23d-b9b5209f3636-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703635 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:09 crc kubenswrapper[4867]: I1212 08:19:09.703649 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/17f2c606-4920-493b-a23d-b9b5209f3636-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.136846 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vpzt4" event={"ID":"17f2c606-4920-493b-a23d-b9b5209f3636","Type":"ContainerDied","Data":"9f26411807f6430e25baa118e3e67ca723540e369aca7a4aaa1f69cfd151b852"} Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.136914 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vpzt4" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.136918 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f26411807f6430e25baa118e3e67ca723540e369aca7a4aaa1f69cfd151b852" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.484474 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:19:10 crc kubenswrapper[4867]: E1212 08:19:10.484842 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" containerName="cinder-db-sync" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.484856 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" containerName="cinder-db-sync" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.485024 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" containerName="cinder-db-sync" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.486616 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.502911 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.621714 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.621898 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.622025 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2qt\" (UniqueName: \"kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.622158 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.622256 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.645837 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.647571 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.649822 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.650800 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.650906 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nb7t6" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.655141 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.661833 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.725041 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.725554 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.725806 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.725949 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.726144 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2qt\" (UniqueName: \"kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.726483 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.726397 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.726904 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.727114 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.745754 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2qt\" (UniqueName: \"kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt\") pod \"dnsmasq-dns-5884c65cf-k9bnc\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.821142 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831642 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831719 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831748 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831765 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831802 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831856 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.831895 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfg7m\" (UniqueName: \"kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943550 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943846 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfg7m\" (UniqueName: \"kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943890 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943921 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943948 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.943962 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.944003 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.949844 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.949922 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.956528 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.956899 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.961769 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.967406 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:10 crc kubenswrapper[4867]: I1212 08:19:10.987818 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfg7m\" (UniqueName: \"kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m\") pod \"cinder-api-0\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " pod="openstack/cinder-api-0" Dec 12 08:19:11 crc kubenswrapper[4867]: I1212 08:19:11.271874 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:11 crc kubenswrapper[4867]: I1212 08:19:11.347598 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:19:11 crc kubenswrapper[4867]: W1212 08:19:11.354147 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd741ed2b_868d_4a97_bde9_aa5f548f9e45.slice/crio-477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae WatchSource:0}: Error finding container 477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae: Status 404 returned error can't find the container with id 477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae Dec 12 08:19:11 crc kubenswrapper[4867]: I1212 08:19:11.713540 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:11 crc kubenswrapper[4867]: W1212 08:19:11.716505 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6558b1d0_f57c_4869_b7e1_5ce82fb19c80.slice/crio-4e8a58e4ca163fb94cde4fe4a4c242c35bfb35f68741e94013a7e40e51423132 WatchSource:0}: Error finding container 4e8a58e4ca163fb94cde4fe4a4c242c35bfb35f68741e94013a7e40e51423132: Status 404 returned error can't find the container with id 4e8a58e4ca163fb94cde4fe4a4c242c35bfb35f68741e94013a7e40e51423132 Dec 12 08:19:12 crc kubenswrapper[4867]: I1212 08:19:12.156695 4867 generic.go:334] "Generic (PLEG): container finished" podID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerID="2868dd45bea43f598497baea5c30c24ca76d1a19d4a00a7dd491dc9990cedf06" exitCode=0 Dec 12 08:19:12 crc kubenswrapper[4867]: I1212 08:19:12.156968 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" event={"ID":"d741ed2b-868d-4a97-bde9-aa5f548f9e45","Type":"ContainerDied","Data":"2868dd45bea43f598497baea5c30c24ca76d1a19d4a00a7dd491dc9990cedf06"} Dec 12 08:19:12 crc kubenswrapper[4867]: I1212 08:19:12.156994 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" event={"ID":"d741ed2b-868d-4a97-bde9-aa5f548f9e45","Type":"ContainerStarted","Data":"477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae"} Dec 12 08:19:12 crc kubenswrapper[4867]: I1212 08:19:12.162327 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerStarted","Data":"4e8a58e4ca163fb94cde4fe4a4c242c35bfb35f68741e94013a7e40e51423132"} Dec 12 08:19:12 crc kubenswrapper[4867]: I1212 08:19:12.677721 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.174289 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" event={"ID":"d741ed2b-868d-4a97-bde9-aa5f548f9e45","Type":"ContainerStarted","Data":"73cd0d4fc21d8b40133a343b64467b830c47f69e8b79463d7de44050ee16cf0c"} Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.174671 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.176009 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerStarted","Data":"c6abf429593b06431469504eca6629c9b329e55776e0715a2e3f40f92fd10cde"} Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.176048 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerStarted","Data":"67c0483e3a93aae26b3f2933bd230cc6e51eee9e450669b18d51f930fcef1717"} Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.176126 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api-log" containerID="cri-o://67c0483e3a93aae26b3f2933bd230cc6e51eee9e450669b18d51f930fcef1717" gracePeriod=30 Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.176175 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.176217 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api" containerID="cri-o://c6abf429593b06431469504eca6629c9b329e55776e0715a2e3f40f92fd10cde" gracePeriod=30 Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.197921 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" podStartSLOduration=3.197897601 podStartE2EDuration="3.197897601s" podCreationTimestamp="2025-12-12 08:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:19:13.19503207 +0000 UTC m=+5440.766413339" watchObservedRunningTime="2025-12-12 08:19:13.197897601 +0000 UTC m=+5440.769278870" Dec 12 08:19:13 crc kubenswrapper[4867]: I1212 08:19:13.222855 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.222837333 podStartE2EDuration="3.222837333s" podCreationTimestamp="2025-12-12 08:19:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:19:13.213473133 +0000 UTC m=+5440.784854402" watchObservedRunningTime="2025-12-12 08:19:13.222837333 +0000 UTC m=+5440.794218602" Dec 12 08:19:14 crc kubenswrapper[4867]: I1212 08:19:14.190629 4867 generic.go:334] "Generic (PLEG): container finished" podID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerID="67c0483e3a93aae26b3f2933bd230cc6e51eee9e450669b18d51f930fcef1717" exitCode=143 Dec 12 08:19:14 crc kubenswrapper[4867]: I1212 08:19:14.190717 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerDied","Data":"67c0483e3a93aae26b3f2933bd230cc6e51eee9e450669b18d51f930fcef1717"} Dec 12 08:19:14 crc kubenswrapper[4867]: I1212 08:19:14.838810 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:19:14 crc kubenswrapper[4867]: E1212 08:19:14.839067 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:19:20 crc kubenswrapper[4867]: I1212 08:19:20.823396 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:19:20 crc kubenswrapper[4867]: I1212 08:19:20.931207 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:19:20 crc kubenswrapper[4867]: I1212 08:19:20.931516 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="dnsmasq-dns" containerID="cri-o://d38185e32cb85a09dd8e447653f266d816a91a8996acc81ea5e31e825dba6291" gracePeriod=10 Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.272474 4867 generic.go:334] "Generic (PLEG): container finished" podID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerID="d38185e32cb85a09dd8e447653f266d816a91a8996acc81ea5e31e825dba6291" exitCode=0 Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.272582 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" event={"ID":"4a869be7-fa0e-4b99-8497-1e48c48acbff","Type":"ContainerDied","Data":"d38185e32cb85a09dd8e447653f266d816a91a8996acc81ea5e31e825dba6291"} Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.512695 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.658371 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc\") pod \"4a869be7-fa0e-4b99-8497-1e48c48acbff\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.658485 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config\") pod \"4a869be7-fa0e-4b99-8497-1e48c48acbff\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.658539 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb\") pod \"4a869be7-fa0e-4b99-8497-1e48c48acbff\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.658614 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb\") pod \"4a869be7-fa0e-4b99-8497-1e48c48acbff\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.658655 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6jr5\" (UniqueName: \"kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5\") pod \"4a869be7-fa0e-4b99-8497-1e48c48acbff\" (UID: \"4a869be7-fa0e-4b99-8497-1e48c48acbff\") " Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.682494 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5" (OuterVolumeSpecName: "kube-api-access-m6jr5") pod "4a869be7-fa0e-4b99-8497-1e48c48acbff" (UID: "4a869be7-fa0e-4b99-8497-1e48c48acbff"). InnerVolumeSpecName "kube-api-access-m6jr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.707132 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4a869be7-fa0e-4b99-8497-1e48c48acbff" (UID: "4a869be7-fa0e-4b99-8497-1e48c48acbff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.708716 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config" (OuterVolumeSpecName: "config") pod "4a869be7-fa0e-4b99-8497-1e48c48acbff" (UID: "4a869be7-fa0e-4b99-8497-1e48c48acbff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.709781 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4a869be7-fa0e-4b99-8497-1e48c48acbff" (UID: "4a869be7-fa0e-4b99-8497-1e48c48acbff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.715189 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4a869be7-fa0e-4b99-8497-1e48c48acbff" (UID: "4a869be7-fa0e-4b99-8497-1e48c48acbff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.760597 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.760634 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.760644 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.760657 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a869be7-fa0e-4b99-8497-1e48c48acbff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:22 crc kubenswrapper[4867]: I1212 08:19:22.760667 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6jr5\" (UniqueName: \"kubernetes.io/projected/4a869be7-fa0e-4b99-8497-1e48c48acbff-kube-api-access-m6jr5\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.283943 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" event={"ID":"4a869be7-fa0e-4b99-8497-1e48c48acbff","Type":"ContainerDied","Data":"59b0dd7f055a999cf6bf94a297ea07794fa6a9c3e4f28dda673e6b172e7e7621"} Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.284250 4867 scope.go:117] "RemoveContainer" containerID="d38185e32cb85a09dd8e447653f266d816a91a8996acc81ea5e31e825dba6291" Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.284390 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb578ccbc-9pmn9" Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.315830 4867 scope.go:117] "RemoveContainer" containerID="b241a464b00969263db9561d612d8fefa8e4c98024d093e712a197461c20da2d" Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.316078 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.325937 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cb578ccbc-9pmn9"] Dec 12 08:19:23 crc kubenswrapper[4867]: I1212 08:19:23.482271 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 12 08:19:24 crc kubenswrapper[4867]: I1212 08:19:24.921886 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" path="/var/lib/kubelet/pods/4a869be7-fa0e-4b99-8497-1e48c48acbff/volumes" Dec 12 08:19:26 crc kubenswrapper[4867]: I1212 08:19:26.843347 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:19:26 crc kubenswrapper[4867]: E1212 08:19:26.844160 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:19:38 crc kubenswrapper[4867]: I1212 08:19:38.838616 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:19:38 crc kubenswrapper[4867]: E1212 08:19:38.839395 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.472146 4867 generic.go:334] "Generic (PLEG): container finished" podID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerID="c6abf429593b06431469504eca6629c9b329e55776e0715a2e3f40f92fd10cde" exitCode=137 Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.472318 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerDied","Data":"c6abf429593b06431469504eca6629c9b329e55776e0715a2e3f40f92fd10cde"} Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.622280 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764627 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764699 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764848 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764898 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfg7m\" (UniqueName: \"kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764920 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764984 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.765025 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data\") pod \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\" (UID: \"6558b1d0-f57c-4869-b7e1-5ce82fb19c80\") " Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.764879 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.765421 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.766140 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs" (OuterVolumeSpecName: "logs") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.769998 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts" (OuterVolumeSpecName: "scripts") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.772160 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.794017 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m" (OuterVolumeSpecName: "kube-api-access-jfg7m") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "kube-api-access-jfg7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.804299 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.818471 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data" (OuterVolumeSpecName: "config-data") pod "6558b1d0-f57c-4869-b7e1-5ce82fb19c80" (UID: "6558b1d0-f57c-4869-b7e1-5ce82fb19c80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866759 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866793 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfg7m\" (UniqueName: \"kubernetes.io/projected/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-kube-api-access-jfg7m\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866806 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866815 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866824 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:43 crc kubenswrapper[4867]: I1212 08:19:43.866864 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6558b1d0-f57c-4869-b7e1-5ce82fb19c80-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.482806 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6558b1d0-f57c-4869-b7e1-5ce82fb19c80","Type":"ContainerDied","Data":"4e8a58e4ca163fb94cde4fe4a4c242c35bfb35f68741e94013a7e40e51423132"} Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.482863 4867 scope.go:117] "RemoveContainer" containerID="c6abf429593b06431469504eca6629c9b329e55776e0715a2e3f40f92fd10cde" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.482869 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.506877 4867 scope.go:117] "RemoveContainer" containerID="67c0483e3a93aae26b3f2933bd230cc6e51eee9e450669b18d51f930fcef1717" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.520777 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.536676 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.545603 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:44 crc kubenswrapper[4867]: E1212 08:19:44.546022 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="dnsmasq-dns" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546047 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="dnsmasq-dns" Dec 12 08:19:44 crc kubenswrapper[4867]: E1212 08:19:44.546068 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546076 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api" Dec 12 08:19:44 crc kubenswrapper[4867]: E1212 08:19:44.546102 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="init" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546110 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="init" Dec 12 08:19:44 crc kubenswrapper[4867]: E1212 08:19:44.546141 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api-log" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546149 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api-log" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546378 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546392 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" containerName="cinder-api-log" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.546416 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a869be7-fa0e-4b99-8497-1e48c48acbff" containerName="dnsmasq-dns" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.547621 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553044 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553299 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553442 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553860 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553997 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nb7t6" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.553873 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.558146 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.683899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.683955 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.683998 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkntk\" (UniqueName: \"kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684294 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684338 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684378 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684492 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.684550 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786078 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786202 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786259 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786285 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkntk\" (UniqueName: \"kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786307 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786306 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786341 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786846 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786920 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.786979 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.787550 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.790833 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.791018 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.791688 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.792094 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.792778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.793196 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.805910 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkntk\" (UniqueName: \"kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk\") pod \"cinder-api-0\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " pod="openstack/cinder-api-0" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.852884 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6558b1d0-f57c-4869-b7e1-5ce82fb19c80" path="/var/lib/kubelet/pods/6558b1d0-f57c-4869-b7e1-5ce82fb19c80/volumes" Dec 12 08:19:44 crc kubenswrapper[4867]: I1212 08:19:44.870561 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:19:45 crc kubenswrapper[4867]: W1212 08:19:45.362402 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ae2b42d_f22a_47a4_9baf_a1c3c16caf72.slice/crio-a2b97b1c655785a7adc116da4e7d711a1f52c49a687d5bbfe13cea76a82745c8 WatchSource:0}: Error finding container a2b97b1c655785a7adc116da4e7d711a1f52c49a687d5bbfe13cea76a82745c8: Status 404 returned error can't find the container with id a2b97b1c655785a7adc116da4e7d711a1f52c49a687d5bbfe13cea76a82745c8 Dec 12 08:19:45 crc kubenswrapper[4867]: I1212 08:19:45.363589 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:19:45 crc kubenswrapper[4867]: I1212 08:19:45.495654 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerStarted","Data":"a2b97b1c655785a7adc116da4e7d711a1f52c49a687d5bbfe13cea76a82745c8"} Dec 12 08:19:46 crc kubenswrapper[4867]: I1212 08:19:46.509693 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerStarted","Data":"811fa362b4835de2fc9f9a48f949fac5117820479aed33db5e613b5750b52a95"} Dec 12 08:19:46 crc kubenswrapper[4867]: I1212 08:19:46.510255 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 12 08:19:46 crc kubenswrapper[4867]: I1212 08:19:46.510274 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerStarted","Data":"1d81a4189f582ed46cabddbe3d81d5f0b328c942a681d0a74833c43adcc87fb0"} Dec 12 08:19:46 crc kubenswrapper[4867]: I1212 08:19:46.537551 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.5375305360000002 podStartE2EDuration="2.537530536s" podCreationTimestamp="2025-12-12 08:19:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:19:46.53202253 +0000 UTC m=+5474.103403819" watchObservedRunningTime="2025-12-12 08:19:46.537530536 +0000 UTC m=+5474.108911815" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.593743 4867 scope.go:117] "RemoveContainer" containerID="008070a9255f964ecb06601f113844333f4d790b3bc7119122f33b9648090e9c" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.621883 4867 scope.go:117] "RemoveContainer" containerID="eb9a1029e25d39c84f7a191d1a9ced6a9c9703852deb7449b67da9801e0a227a" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.673105 4867 scope.go:117] "RemoveContainer" containerID="a48842721da335fda06794b72126d6dac1510a9b06123f879baad397e6cde37d" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.711436 4867 scope.go:117] "RemoveContainer" containerID="87a8dd351b2ace9a29ffd52442b4dce7dff5ec54d3828f0fbaf4f2be1fdaaf0b" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.735846 4867 scope.go:117] "RemoveContainer" containerID="7525e1e93a72e8fb116d175998e511e0b9a6955a80e82c3df82426e39d1e7c4e" Dec 12 08:19:53 crc kubenswrapper[4867]: I1212 08:19:53.839879 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:19:53 crc kubenswrapper[4867]: E1212 08:19:53.840529 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:19:56 crc kubenswrapper[4867]: I1212 08:19:56.798602 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 12 08:20:07 crc kubenswrapper[4867]: I1212 08:20:07.837681 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:20:07 crc kubenswrapper[4867]: E1212 08:20:07.838417 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.837838 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:20:18 crc kubenswrapper[4867]: E1212 08:20:18.838771 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.870563 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.872335 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.873942 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.885109 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.965794 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdkmj\" (UniqueName: \"kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.966106 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.966223 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.966377 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.966482 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:18 crc kubenswrapper[4867]: I1212 08:20:18.966581 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.068438 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdkmj\" (UniqueName: \"kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.068546 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.068577 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.069631 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.069680 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.069733 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.069922 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.075528 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.078748 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.079170 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.079543 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.094195 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdkmj\" (UniqueName: \"kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj\") pod \"cinder-scheduler-0\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.189355 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.647407 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.659631 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:20:19 crc kubenswrapper[4867]: I1212 08:20:19.799077 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerStarted","Data":"218d3c4b1c0bef43963932b6d49eafba862ad2f402ee5767a52170f61a9f7bc2"} Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.409748 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.410399 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api" containerID="cri-o://811fa362b4835de2fc9f9a48f949fac5117820479aed33db5e613b5750b52a95" gracePeriod=30 Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.410316 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api-log" containerID="cri-o://1d81a4189f582ed46cabddbe3d81d5f0b328c942a681d0a74833c43adcc87fb0" gracePeriod=30 Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.814424 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerStarted","Data":"fa1b1304e89071ea3ed0d097dae918aef2601e084bb14b6155146a6347c6307c"} Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.817033 4867 generic.go:334] "Generic (PLEG): container finished" podID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerID="1d81a4189f582ed46cabddbe3d81d5f0b328c942a681d0a74833c43adcc87fb0" exitCode=143 Dec 12 08:20:20 crc kubenswrapper[4867]: I1212 08:20:20.817086 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerDied","Data":"1d81a4189f582ed46cabddbe3d81d5f0b328c942a681d0a74833c43adcc87fb0"} Dec 12 08:20:21 crc kubenswrapper[4867]: I1212 08:20:21.831764 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerStarted","Data":"4a73e1d8e8cab1ba58cf76a887a5d2b85342f741af0af0c13bbb405e30d1774d"} Dec 12 08:20:23 crc kubenswrapper[4867]: I1212 08:20:23.855141 4867 generic.go:334] "Generic (PLEG): container finished" podID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerID="811fa362b4835de2fc9f9a48f949fac5117820479aed33db5e613b5750b52a95" exitCode=0 Dec 12 08:20:23 crc kubenswrapper[4867]: I1212 08:20:23.855215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerDied","Data":"811fa362b4835de2fc9f9a48f949fac5117820479aed33db5e613b5750b52a95"} Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.021272 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.038484 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.768255098 podStartE2EDuration="6.03846183s" podCreationTimestamp="2025-12-12 08:20:18 +0000 UTC" firstStartedPulling="2025-12-12 08:20:19.659364706 +0000 UTC m=+5507.230745975" lastFinishedPulling="2025-12-12 08:20:19.929571438 +0000 UTC m=+5507.500952707" observedRunningTime="2025-12-12 08:20:21.862181432 +0000 UTC m=+5509.433562721" watchObservedRunningTime="2025-12-12 08:20:24.03846183 +0000 UTC m=+5511.609843089" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170319 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170384 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170440 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170468 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170556 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkntk\" (UniqueName: \"kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170601 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170635 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.170678 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts\") pod \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\" (UID: \"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72\") " Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.171196 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs" (OuterVolumeSpecName: "logs") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.171309 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.176485 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts" (OuterVolumeSpecName: "scripts") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.181191 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.182431 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk" (OuterVolumeSpecName: "kube-api-access-mkntk") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "kube-api-access-mkntk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.190146 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.215400 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.239085 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.262969 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data" (OuterVolumeSpecName: "config-data") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272846 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272880 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272891 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272899 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkntk\" (UniqueName: \"kubernetes.io/projected/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-kube-api-access-mkntk\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272908 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272918 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272929 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.272937 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.290786 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" (UID: "2ae2b42d-f22a-47a4-9baf-a1c3c16caf72"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.374405 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.870215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2ae2b42d-f22a-47a4-9baf-a1c3c16caf72","Type":"ContainerDied","Data":"a2b97b1c655785a7adc116da4e7d711a1f52c49a687d5bbfe13cea76a82745c8"} Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.870310 4867 scope.go:117] "RemoveContainer" containerID="811fa362b4835de2fc9f9a48f949fac5117820479aed33db5e613b5750b52a95" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.870338 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.893191 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.901869 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.906081 4867 scope.go:117] "RemoveContainer" containerID="1d81a4189f582ed46cabddbe3d81d5f0b328c942a681d0a74833c43adcc87fb0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.925634 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:24 crc kubenswrapper[4867]: E1212 08:20:24.926007 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.926020 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api" Dec 12 08:20:24 crc kubenswrapper[4867]: E1212 08:20:24.926035 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api-log" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.926041 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api-log" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.926273 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.926298 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" containerName="cinder-api-log" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.927269 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.932939 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.934607 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.935315 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.952828 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985266 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985344 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985419 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb57g\" (UniqueName: \"kubernetes.io/projected/a7368375-0f30-4a7f-9905-7786ed91be0b-kube-api-access-mb57g\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985483 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7368375-0f30-4a7f-9905-7786ed91be0b-logs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985595 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985669 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7368375-0f30-4a7f-9905-7786ed91be0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985736 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985810 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-scripts\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:24 crc kubenswrapper[4867]: I1212 08:20:24.985880 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087381 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7368375-0f30-4a7f-9905-7786ed91be0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087457 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087505 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-scripts\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087514 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a7368375-0f30-4a7f-9905-7786ed91be0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087553 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087609 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087642 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087675 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb57g\" (UniqueName: \"kubernetes.io/projected/a7368375-0f30-4a7f-9905-7786ed91be0b-kube-api-access-mb57g\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087719 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7368375-0f30-4a7f-9905-7786ed91be0b-logs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.087741 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.088475 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7368375-0f30-4a7f-9905-7786ed91be0b-logs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.092275 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-scripts\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.092967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.093427 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.093793 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.095679 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-config-data\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.098320 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7368375-0f30-4a7f-9905-7786ed91be0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.104733 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb57g\" (UniqueName: \"kubernetes.io/projected/a7368375-0f30-4a7f-9905-7786ed91be0b-kube-api-access-mb57g\") pod \"cinder-api-0\" (UID: \"a7368375-0f30-4a7f-9905-7786ed91be0b\") " pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.248196 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.714046 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 12 08:20:25 crc kubenswrapper[4867]: I1212 08:20:25.883400 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a7368375-0f30-4a7f-9905-7786ed91be0b","Type":"ContainerStarted","Data":"f49b77e5518a7f0cb13bec53c40334a107922eb08c7920d0fd8c7c186496f37e"} Dec 12 08:20:26 crc kubenswrapper[4867]: I1212 08:20:26.849453 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae2b42d-f22a-47a4-9baf-a1c3c16caf72" path="/var/lib/kubelet/pods/2ae2b42d-f22a-47a4-9baf-a1c3c16caf72/volumes" Dec 12 08:20:26 crc kubenswrapper[4867]: I1212 08:20:26.894413 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a7368375-0f30-4a7f-9905-7786ed91be0b","Type":"ContainerStarted","Data":"3b4c27771b88863ae01a0473bb6993a833a819e01261c95bddca6e15251510a6"} Dec 12 08:20:26 crc kubenswrapper[4867]: I1212 08:20:26.894457 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a7368375-0f30-4a7f-9905-7786ed91be0b","Type":"ContainerStarted","Data":"5fbf3475838094d061302454d849a12e9c3e0b9db5eff5bedd977177b75d2d6e"} Dec 12 08:20:26 crc kubenswrapper[4867]: I1212 08:20:26.895616 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 12 08:20:26 crc kubenswrapper[4867]: I1212 08:20:26.921718 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.921696463 podStartE2EDuration="2.921696463s" podCreationTimestamp="2025-12-12 08:20:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:20:26.910932536 +0000 UTC m=+5514.482313795" watchObservedRunningTime="2025-12-12 08:20:26.921696463 +0000 UTC m=+5514.493077732" Dec 12 08:20:29 crc kubenswrapper[4867]: I1212 08:20:29.445372 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 12 08:20:29 crc kubenswrapper[4867]: I1212 08:20:29.519141 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:29 crc kubenswrapper[4867]: I1212 08:20:29.923984 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="cinder-scheduler" containerID="cri-o://fa1b1304e89071ea3ed0d097dae918aef2601e084bb14b6155146a6347c6307c" gracePeriod=30 Dec 12 08:20:29 crc kubenswrapper[4867]: I1212 08:20:29.924210 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="probe" containerID="cri-o://4a73e1d8e8cab1ba58cf76a887a5d2b85342f741af0af0c13bbb405e30d1774d" gracePeriod=30 Dec 12 08:20:30 crc kubenswrapper[4867]: I1212 08:20:30.938934 4867 generic.go:334] "Generic (PLEG): container finished" podID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerID="4a73e1d8e8cab1ba58cf76a887a5d2b85342f741af0af0c13bbb405e30d1774d" exitCode=0 Dec 12 08:20:30 crc kubenswrapper[4867]: I1212 08:20:30.939006 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerDied","Data":"4a73e1d8e8cab1ba58cf76a887a5d2b85342f741af0af0c13bbb405e30d1774d"} Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.003591 4867 generic.go:334] "Generic (PLEG): container finished" podID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerID="fa1b1304e89071ea3ed0d097dae918aef2601e084bb14b6155146a6347c6307c" exitCode=0 Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.003906 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerDied","Data":"fa1b1304e89071ea3ed0d097dae918aef2601e084bb14b6155146a6347c6307c"} Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.047862 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.137652 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.137749 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.137889 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdkmj\" (UniqueName: \"kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.137936 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.138031 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.138153 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id\") pod \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\" (UID: \"3c06b4ba-ab77-4960-aa66-4f61f939ee25\") " Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.138687 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.147173 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts" (OuterVolumeSpecName: "scripts") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.147644 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj" (OuterVolumeSpecName: "kube-api-access-sdkmj") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "kube-api-access-sdkmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.150027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.201152 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.240517 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.240554 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdkmj\" (UniqueName: \"kubernetes.io/projected/3c06b4ba-ab77-4960-aa66-4f61f939ee25-kube-api-access-sdkmj\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.240568 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.240579 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.240587 4867 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c06b4ba-ab77-4960-aa66-4f61f939ee25-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.242589 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data" (OuterVolumeSpecName: "config-data") pod "3c06b4ba-ab77-4960-aa66-4f61f939ee25" (UID: "3c06b4ba-ab77-4960-aa66-4f61f939ee25"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.342173 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c06b4ba-ab77-4960-aa66-4f61f939ee25-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:32 crc kubenswrapper[4867]: I1212 08:20:32.839746 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:20:32 crc kubenswrapper[4867]: E1212 08:20:32.840140 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.017986 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3c06b4ba-ab77-4960-aa66-4f61f939ee25","Type":"ContainerDied","Data":"218d3c4b1c0bef43963932b6d49eafba862ad2f402ee5767a52170f61a9f7bc2"} Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.018285 4867 scope.go:117] "RemoveContainer" containerID="4a73e1d8e8cab1ba58cf76a887a5d2b85342f741af0af0c13bbb405e30d1774d" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.018421 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.051167 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.065529 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.069415 4867 scope.go:117] "RemoveContainer" containerID="fa1b1304e89071ea3ed0d097dae918aef2601e084bb14b6155146a6347c6307c" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.084897 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:33 crc kubenswrapper[4867]: E1212 08:20:33.085574 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="cinder-scheduler" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.085603 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="cinder-scheduler" Dec 12 08:20:33 crc kubenswrapper[4867]: E1212 08:20:33.085628 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="probe" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.085642 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="probe" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.085918 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="cinder-scheduler" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.085945 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" containerName="probe" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.088105 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.091086 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.098641 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156204 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf535deb-d3aa-48ae-9c8d-190d12647967-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156401 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156460 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrdc\" (UniqueName: \"kubernetes.io/projected/bf535deb-d3aa-48ae-9c8d-190d12647967-kube-api-access-gxrdc\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156748 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156800 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.156901 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.258335 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf535deb-d3aa-48ae-9c8d-190d12647967-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.258674 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.258807 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrdc\" (UniqueName: \"kubernetes.io/projected/bf535deb-d3aa-48ae-9c8d-190d12647967-kube-api-access-gxrdc\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.258988 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.259101 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.259252 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.258458 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bf535deb-d3aa-48ae-9c8d-190d12647967-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.263346 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-scripts\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.263831 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.274889 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.275779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf535deb-d3aa-48ae-9c8d-190d12647967-config-data\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.277973 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrdc\" (UniqueName: \"kubernetes.io/projected/bf535deb-d3aa-48ae-9c8d-190d12647967-kube-api-access-gxrdc\") pod \"cinder-scheduler-0\" (UID: \"bf535deb-d3aa-48ae-9c8d-190d12647967\") " pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.416687 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 12 08:20:33 crc kubenswrapper[4867]: I1212 08:20:33.861627 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 12 08:20:34 crc kubenswrapper[4867]: I1212 08:20:34.033683 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf535deb-d3aa-48ae-9c8d-190d12647967","Type":"ContainerStarted","Data":"61d5ccc23760b7e3507ccd2598a662be93fe387bf057473c57b5e2dd36d91d5f"} Dec 12 08:20:34 crc kubenswrapper[4867]: I1212 08:20:34.856105 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c06b4ba-ab77-4960-aa66-4f61f939ee25" path="/var/lib/kubelet/pods/3c06b4ba-ab77-4960-aa66-4f61f939ee25/volumes" Dec 12 08:20:35 crc kubenswrapper[4867]: I1212 08:20:35.043842 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf535deb-d3aa-48ae-9c8d-190d12647967","Type":"ContainerStarted","Data":"b5db6f6fe80177120309bfc12a20a5335c7543b0c48702eed94614904652bc5c"} Dec 12 08:20:35 crc kubenswrapper[4867]: I1212 08:20:35.043887 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bf535deb-d3aa-48ae-9c8d-190d12647967","Type":"ContainerStarted","Data":"8dc82c0935c513ba3365ec4b8fb2f28f7b75cd4eb5fb73df5f7f001a68492929"} Dec 12 08:20:35 crc kubenswrapper[4867]: I1212 08:20:35.067142 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.067121912 podStartE2EDuration="2.067121912s" podCreationTimestamp="2025-12-12 08:20:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:20:35.062886277 +0000 UTC m=+5522.634267546" watchObservedRunningTime="2025-12-12 08:20:35.067121912 +0000 UTC m=+5522.638503181" Dec 12 08:20:37 crc kubenswrapper[4867]: I1212 08:20:37.216370 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 12 08:20:38 crc kubenswrapper[4867]: I1212 08:20:38.417135 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 12 08:20:43 crc kubenswrapper[4867]: I1212 08:20:43.622266 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 12 08:20:43 crc kubenswrapper[4867]: I1212 08:20:43.838460 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:20:43 crc kubenswrapper[4867]: E1212 08:20:43.839201 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.054003 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ksqq5"] Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.055085 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.072872 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ksqq5"] Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.157891 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4768-account-create-update-rcgbk"] Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.159454 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.161860 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.227374 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2zvc\" (UniqueName: \"kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.227505 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.234468 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4768-account-create-update-rcgbk"] Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.329454 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.329562 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sztf\" (UniqueName: \"kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.329595 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2zvc\" (UniqueName: \"kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.329686 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.330478 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.351672 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2zvc\" (UniqueName: \"kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc\") pod \"glance-db-create-ksqq5\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.375355 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.431035 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sztf\" (UniqueName: \"kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.431209 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.431939 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.450753 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sztf\" (UniqueName: \"kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf\") pod \"glance-4768-account-create-update-rcgbk\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.561126 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.883518 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4768-account-create-update-rcgbk"] Dec 12 08:20:44 crc kubenswrapper[4867]: I1212 08:20:44.892730 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ksqq5"] Dec 12 08:20:45 crc kubenswrapper[4867]: I1212 08:20:45.121762 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ksqq5" event={"ID":"22edfff0-ed98-4a5d-b74b-88c8bc029cb7","Type":"ContainerStarted","Data":"c0f4947dce2179f62d2e10f1a1abf059f50e01ee5945524fb71ae23c7d314829"} Dec 12 08:20:45 crc kubenswrapper[4867]: I1212 08:20:45.123735 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4768-account-create-update-rcgbk" event={"ID":"65ef59d9-3ad2-4652-9ed3-69f1798c89d0","Type":"ContainerStarted","Data":"6c64dcbe610111b48ce85c7fc72858fb9e0fa8073d9214110a2bf4193f848867"} Dec 12 08:20:46 crc kubenswrapper[4867]: I1212 08:20:46.136891 4867 generic.go:334] "Generic (PLEG): container finished" podID="65ef59d9-3ad2-4652-9ed3-69f1798c89d0" containerID="2101f0952f228d0909a976df96b4c591d3c99076d3390137125da4e5a04e04a1" exitCode=0 Dec 12 08:20:46 crc kubenswrapper[4867]: I1212 08:20:46.136982 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4768-account-create-update-rcgbk" event={"ID":"65ef59d9-3ad2-4652-9ed3-69f1798c89d0","Type":"ContainerDied","Data":"2101f0952f228d0909a976df96b4c591d3c99076d3390137125da4e5a04e04a1"} Dec 12 08:20:46 crc kubenswrapper[4867]: I1212 08:20:46.143281 4867 generic.go:334] "Generic (PLEG): container finished" podID="22edfff0-ed98-4a5d-b74b-88c8bc029cb7" containerID="a6ee51b17db7d5155140ba45a07c9a60837aa49608c73b8eb7204e2444fd7a64" exitCode=0 Dec 12 08:20:46 crc kubenswrapper[4867]: I1212 08:20:46.143341 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ksqq5" event={"ID":"22edfff0-ed98-4a5d-b74b-88c8bc029cb7","Type":"ContainerDied","Data":"a6ee51b17db7d5155140ba45a07c9a60837aa49608c73b8eb7204e2444fd7a64"} Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.535346 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.547486 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.695105 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts\") pod \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.695368 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2zvc\" (UniqueName: \"kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc\") pod \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\" (UID: \"22edfff0-ed98-4a5d-b74b-88c8bc029cb7\") " Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.695394 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sztf\" (UniqueName: \"kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf\") pod \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.695423 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts\") pod \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\" (UID: \"65ef59d9-3ad2-4652-9ed3-69f1798c89d0\") " Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.696027 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65ef59d9-3ad2-4652-9ed3-69f1798c89d0" (UID: "65ef59d9-3ad2-4652-9ed3-69f1798c89d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.696445 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22edfff0-ed98-4a5d-b74b-88c8bc029cb7" (UID: "22edfff0-ed98-4a5d-b74b-88c8bc029cb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.700378 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc" (OuterVolumeSpecName: "kube-api-access-h2zvc") pod "22edfff0-ed98-4a5d-b74b-88c8bc029cb7" (UID: "22edfff0-ed98-4a5d-b74b-88c8bc029cb7"). InnerVolumeSpecName "kube-api-access-h2zvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.700486 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf" (OuterVolumeSpecName: "kube-api-access-5sztf") pod "65ef59d9-3ad2-4652-9ed3-69f1798c89d0" (UID: "65ef59d9-3ad2-4652-9ed3-69f1798c89d0"). InnerVolumeSpecName "kube-api-access-5sztf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.797457 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2zvc\" (UniqueName: \"kubernetes.io/projected/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-kube-api-access-h2zvc\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.797497 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sztf\" (UniqueName: \"kubernetes.io/projected/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-kube-api-access-5sztf\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.797509 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ef59d9-3ad2-4652-9ed3-69f1798c89d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:47 crc kubenswrapper[4867]: I1212 08:20:47.797519 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22edfff0-ed98-4a5d-b74b-88c8bc029cb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.163922 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4768-account-create-update-rcgbk" event={"ID":"65ef59d9-3ad2-4652-9ed3-69f1798c89d0","Type":"ContainerDied","Data":"6c64dcbe610111b48ce85c7fc72858fb9e0fa8073d9214110a2bf4193f848867"} Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.163959 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c64dcbe610111b48ce85c7fc72858fb9e0fa8073d9214110a2bf4193f848867" Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.163933 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4768-account-create-update-rcgbk" Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.165506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ksqq5" event={"ID":"22edfff0-ed98-4a5d-b74b-88c8bc029cb7","Type":"ContainerDied","Data":"c0f4947dce2179f62d2e10f1a1abf059f50e01ee5945524fb71ae23c7d314829"} Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.165619 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f4947dce2179f62d2e10f1a1abf059f50e01ee5945524fb71ae23c7d314829" Dec 12 08:20:48 crc kubenswrapper[4867]: I1212 08:20:48.165591 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ksqq5" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.382515 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-457vz"] Dec 12 08:20:49 crc kubenswrapper[4867]: E1212 08:20:49.383159 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ef59d9-3ad2-4652-9ed3-69f1798c89d0" containerName="mariadb-account-create-update" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.383173 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ef59d9-3ad2-4652-9ed3-69f1798c89d0" containerName="mariadb-account-create-update" Dec 12 08:20:49 crc kubenswrapper[4867]: E1212 08:20:49.383195 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22edfff0-ed98-4a5d-b74b-88c8bc029cb7" containerName="mariadb-database-create" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.383200 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="22edfff0-ed98-4a5d-b74b-88c8bc029cb7" containerName="mariadb-database-create" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.383397 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="22edfff0-ed98-4a5d-b74b-88c8bc029cb7" containerName="mariadb-database-create" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.383422 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ef59d9-3ad2-4652-9ed3-69f1798c89d0" containerName="mariadb-account-create-update" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.384148 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.386430 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.390743 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zqqmt" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.403586 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-457vz"] Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.528771 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.529168 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.529445 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.529751 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr2pz\" (UniqueName: \"kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.630743 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.630811 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.630841 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.630893 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr2pz\" (UniqueName: \"kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.641083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.641083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.641085 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.647894 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr2pz\" (UniqueName: \"kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz\") pod \"glance-db-sync-457vz\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " pod="openstack/glance-db-sync-457vz" Dec 12 08:20:49 crc kubenswrapper[4867]: I1212 08:20:49.700483 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-457vz" Dec 12 08:20:50 crc kubenswrapper[4867]: I1212 08:20:50.245319 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-457vz"] Dec 12 08:20:51 crc kubenswrapper[4867]: I1212 08:20:51.202893 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-457vz" event={"ID":"fceb31cc-12c6-46ab-85d2-107a5b9696d3","Type":"ContainerStarted","Data":"7334f316dddae0a03d482618ce8f01f2b3b19f50e68f164a9df3e2c5cc523de5"} Dec 12 08:20:56 crc kubenswrapper[4867]: I1212 08:20:56.837873 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:20:56 crc kubenswrapper[4867]: E1212 08:20:56.838601 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:21:07 crc kubenswrapper[4867]: I1212 08:21:07.347574 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-457vz" event={"ID":"fceb31cc-12c6-46ab-85d2-107a5b9696d3","Type":"ContainerStarted","Data":"ea428821739f255818a70b3c565d6503762056608cab40f46c73e1f01bafd3fd"} Dec 12 08:21:07 crc kubenswrapper[4867]: I1212 08:21:07.363346 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-457vz" podStartSLOduration=1.8709494530000002 podStartE2EDuration="18.363322781s" podCreationTimestamp="2025-12-12 08:20:49 +0000 UTC" firstStartedPulling="2025-12-12 08:20:50.259591519 +0000 UTC m=+5537.830972788" lastFinishedPulling="2025-12-12 08:21:06.751964847 +0000 UTC m=+5554.323346116" observedRunningTime="2025-12-12 08:21:07.361650729 +0000 UTC m=+5554.933031988" watchObservedRunningTime="2025-12-12 08:21:07.363322781 +0000 UTC m=+5554.934704090" Dec 12 08:21:11 crc kubenswrapper[4867]: I1212 08:21:11.384384 4867 generic.go:334] "Generic (PLEG): container finished" podID="fceb31cc-12c6-46ab-85d2-107a5b9696d3" containerID="ea428821739f255818a70b3c565d6503762056608cab40f46c73e1f01bafd3fd" exitCode=0 Dec 12 08:21:11 crc kubenswrapper[4867]: I1212 08:21:11.384476 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-457vz" event={"ID":"fceb31cc-12c6-46ab-85d2-107a5b9696d3","Type":"ContainerDied","Data":"ea428821739f255818a70b3c565d6503762056608cab40f46c73e1f01bafd3fd"} Dec 12 08:21:11 crc kubenswrapper[4867]: I1212 08:21:11.838296 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:21:11 crc kubenswrapper[4867]: E1212 08:21:11.838550 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.876569 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-457vz" Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.989779 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle\") pod \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.989877 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data\") pod \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.990026 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr2pz\" (UniqueName: \"kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz\") pod \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.990190 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data\") pod \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\" (UID: \"fceb31cc-12c6-46ab-85d2-107a5b9696d3\") " Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.995465 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fceb31cc-12c6-46ab-85d2-107a5b9696d3" (UID: "fceb31cc-12c6-46ab-85d2-107a5b9696d3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:12 crc kubenswrapper[4867]: I1212 08:21:12.996464 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz" (OuterVolumeSpecName: "kube-api-access-nr2pz") pod "fceb31cc-12c6-46ab-85d2-107a5b9696d3" (UID: "fceb31cc-12c6-46ab-85d2-107a5b9696d3"). InnerVolumeSpecName "kube-api-access-nr2pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.015447 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fceb31cc-12c6-46ab-85d2-107a5b9696d3" (UID: "fceb31cc-12c6-46ab-85d2-107a5b9696d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.038009 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data" (OuterVolumeSpecName: "config-data") pod "fceb31cc-12c6-46ab-85d2-107a5b9696d3" (UID: "fceb31cc-12c6-46ab-85d2-107a5b9696d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.092593 4867 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.092643 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.092659 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fceb31cc-12c6-46ab-85d2-107a5b9696d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.092670 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr2pz\" (UniqueName: \"kubernetes.io/projected/fceb31cc-12c6-46ab-85d2-107a5b9696d3-kube-api-access-nr2pz\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.406695 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-457vz" event={"ID":"fceb31cc-12c6-46ab-85d2-107a5b9696d3","Type":"ContainerDied","Data":"7334f316dddae0a03d482618ce8f01f2b3b19f50e68f164a9df3e2c5cc523de5"} Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.406739 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7334f316dddae0a03d482618ce8f01f2b3b19f50e68f164a9df3e2c5cc523de5" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.406819 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-457vz" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.751547 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:13 crc kubenswrapper[4867]: E1212 08:21:13.752278 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fceb31cc-12c6-46ab-85d2-107a5b9696d3" containerName="glance-db-sync" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.752294 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fceb31cc-12c6-46ab-85d2-107a5b9696d3" containerName="glance-db-sync" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.752479 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fceb31cc-12c6-46ab-85d2-107a5b9696d3" containerName="glance-db-sync" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.753456 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.768136 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.821558 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.823459 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.836595 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.839629 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zqqmt" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.841252 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.854119 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903717 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903777 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903806 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903831 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwgm\" (UniqueName: \"kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903857 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.903942 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.904174 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.904248 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.904290 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74j9g\" (UniqueName: \"kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.904339 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.904467 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.969050 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.971554 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.979717 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 08:21:13 crc kubenswrapper[4867]: I1212 08:21:13.989271 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012161 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012363 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012514 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwgm\" (UniqueName: \"kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012553 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012577 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012685 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012730 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012765 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74j9g\" (UniqueName: \"kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.012804 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.014180 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.014180 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.014917 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.015941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.016217 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.018802 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.029611 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.031595 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.032768 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.034880 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74j9g\" (UniqueName: \"kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g\") pod \"dnsmasq-dns-5f9f666b79-bxtzx\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.051846 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwgm\" (UniqueName: \"kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm\") pod \"glance-default-external-api-0\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.077387 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.114778 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.114891 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dv6l\" (UniqueName: \"kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.114943 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.115028 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.115064 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.115107 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.146923 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.221418 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.221894 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.221948 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.221960 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.222126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.222259 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.222491 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dv6l\" (UniqueName: \"kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.223356 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.229996 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.230073 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.230291 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.243305 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dv6l\" (UniqueName: \"kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l\") pod \"glance-default-internal-api-0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.289903 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:14 crc kubenswrapper[4867]: I1212 08:21:14.571401 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:15 crc kubenswrapper[4867]: W1212 08:21:14.822257 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e079e44_c4cb_4304_a3a6_b7814e4101d1.slice/crio-74590427b0fd4656da8282bdf12498d8d76eec1addf548340d08c860cd02a404 WatchSource:0}: Error finding container 74590427b0fd4656da8282bdf12498d8d76eec1addf548340d08c860cd02a404: Status 404 returned error can't find the container with id 74590427b0fd4656da8282bdf12498d8d76eec1addf548340d08c860cd02a404 Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:14.822696 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:14.924527 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:15 crc kubenswrapper[4867]: W1212 08:21:14.926542 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bf77f96_18df_48d2_a7cc_7954ba9039f0.slice/crio-1c933ef4bb23ae06bc8e8c929c1890e5abe607d0458c8ae13c0b91eda6dc7f07 WatchSource:0}: Error finding container 1c933ef4bb23ae06bc8e8c929c1890e5abe607d0458c8ae13c0b91eda6dc7f07: Status 404 returned error can't find the container with id 1c933ef4bb23ae06bc8e8c929c1890e5abe607d0458c8ae13c0b91eda6dc7f07 Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.095293 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.437148 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerStarted","Data":"1c933ef4bb23ae06bc8e8c929c1890e5abe607d0458c8ae13c0b91eda6dc7f07"} Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.445994 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerStarted","Data":"74590427b0fd4656da8282bdf12498d8d76eec1addf548340d08c860cd02a404"} Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.452824 4867 generic.go:334] "Generic (PLEG): container finished" podID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerID="7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278" exitCode=0 Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.452864 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" event={"ID":"106d309a-9808-4be0-a2ee-6a7522ebc357","Type":"ContainerDied","Data":"7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278"} Dec 12 08:21:15 crc kubenswrapper[4867]: I1212 08:21:15.452892 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" event={"ID":"106d309a-9808-4be0-a2ee-6a7522ebc357","Type":"ContainerStarted","Data":"349d3cf77a9d14150a34090dabf5fe96318317af5edeb159963bd4bfe3c08339"} Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.261075 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.465982 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-log" containerID="cri-o://71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" gracePeriod=30 Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.466132 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-httpd" containerID="cri-o://5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" gracePeriod=30 Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.466298 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerStarted","Data":"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3"} Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.466507 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerStarted","Data":"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9"} Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.468921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" event={"ID":"106d309a-9808-4be0-a2ee-6a7522ebc357","Type":"ContainerStarted","Data":"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862"} Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.469074 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.473032 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerStarted","Data":"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088"} Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.487665 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.48764135 podStartE2EDuration="3.48764135s" podCreationTimestamp="2025-12-12 08:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:16.485405675 +0000 UTC m=+5564.056786944" watchObservedRunningTime="2025-12-12 08:21:16.48764135 +0000 UTC m=+5564.059022619" Dec 12 08:21:16 crc kubenswrapper[4867]: I1212 08:21:16.520026 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" podStartSLOduration=3.520001762 podStartE2EDuration="3.520001762s" podCreationTimestamp="2025-12-12 08:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:16.517242254 +0000 UTC m=+5564.088623533" watchObservedRunningTime="2025-12-12 08:21:16.520001762 +0000 UTC m=+5564.091383051" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.153307 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.305770 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.305974 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306052 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306142 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvwgm\" (UniqueName: \"kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306176 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306201 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs\") pod \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\" (UID: \"3e079e44-c4cb-4304-a3a6-b7814e4101d1\") " Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306581 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.306717 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs" (OuterVolumeSpecName: "logs") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.312462 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts" (OuterVolumeSpecName: "scripts") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.315856 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm" (OuterVolumeSpecName: "kube-api-access-mvwgm") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "kube-api-access-mvwgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.345085 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.380556 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data" (OuterVolumeSpecName: "config-data") pod "3e079e44-c4cb-4304-a3a6-b7814e4101d1" (UID: "3e079e44-c4cb-4304-a3a6-b7814e4101d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408383 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408558 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408574 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvwgm\" (UniqueName: \"kubernetes.io/projected/3e079e44-c4cb-4304-a3a6-b7814e4101d1-kube-api-access-mvwgm\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408584 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408594 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e079e44-c4cb-4304-a3a6-b7814e4101d1-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.408602 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e079e44-c4cb-4304-a3a6-b7814e4101d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.484957 4867 generic.go:334] "Generic (PLEG): container finished" podID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerID="5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" exitCode=143 Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485176 4867 generic.go:334] "Generic (PLEG): container finished" podID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerID="71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" exitCode=143 Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485038 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerDied","Data":"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9"} Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485369 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerDied","Data":"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3"} Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3e079e44-c4cb-4304-a3a6-b7814e4101d1","Type":"ContainerDied","Data":"74590427b0fd4656da8282bdf12498d8d76eec1addf548340d08c860cd02a404"} Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485399 4867 scope.go:117] "RemoveContainer" containerID="5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.485012 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.488418 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerStarted","Data":"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d"} Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.488592 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-log" containerID="cri-o://21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" gracePeriod=30 Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.488607 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-httpd" containerID="cri-o://7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" gracePeriod=30 Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.511388 4867 scope.go:117] "RemoveContainer" containerID="71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.532381 4867 scope.go:117] "RemoveContainer" containerID="5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.535438 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.535413148 podStartE2EDuration="4.535413148s" podCreationTimestamp="2025-12-12 08:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:17.523169644 +0000 UTC m=+5565.094550913" watchObservedRunningTime="2025-12-12 08:21:17.535413148 +0000 UTC m=+5565.106794427" Dec 12 08:21:17 crc kubenswrapper[4867]: E1212 08:21:17.546385 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9\": container with ID starting with 5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9 not found: ID does not exist" containerID="5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.546447 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9"} err="failed to get container status \"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9\": rpc error: code = NotFound desc = could not find container \"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9\": container with ID starting with 5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9 not found: ID does not exist" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.546477 4867 scope.go:117] "RemoveContainer" containerID="71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" Dec 12 08:21:17 crc kubenswrapper[4867]: E1212 08:21:17.546976 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3\": container with ID starting with 71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3 not found: ID does not exist" containerID="71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.547026 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3"} err="failed to get container status \"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3\": rpc error: code = NotFound desc = could not find container \"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3\": container with ID starting with 71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3 not found: ID does not exist" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.547051 4867 scope.go:117] "RemoveContainer" containerID="5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.547537 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9"} err="failed to get container status \"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9\": rpc error: code = NotFound desc = could not find container \"5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9\": container with ID starting with 5397f2d194d7b50a217cd606f5cadece5f073566eab15a53f0db590720ecf6d9 not found: ID does not exist" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.547561 4867 scope.go:117] "RemoveContainer" containerID="71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.547884 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3"} err="failed to get container status \"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3\": rpc error: code = NotFound desc = could not find container \"71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3\": container with ID starting with 71df0dd8e78c817a23d64ac09e825550108f87667f41b95414b39d92a9b3e8f3 not found: ID does not exist" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.554719 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.569922 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.582745 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:17 crc kubenswrapper[4867]: E1212 08:21:17.583183 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-log" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.583202 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-log" Dec 12 08:21:17 crc kubenswrapper[4867]: E1212 08:21:17.583240 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-httpd" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.583247 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-httpd" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.583435 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-httpd" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.583457 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" containerName="glance-log" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.584425 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.587158 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.587682 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.594616 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714322 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714370 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714414 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6b7r\" (UniqueName: \"kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714578 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714758 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.714920 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.715059 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816721 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816780 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816828 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6b7r\" (UniqueName: \"kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816858 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816896 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.816939 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.817075 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.817465 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.817583 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.821290 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.822299 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.822844 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.824854 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.837930 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6b7r\" (UniqueName: \"kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r\") pod \"glance-default-external-api-0\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " pod="openstack/glance-default-external-api-0" Dec 12 08:21:17 crc kubenswrapper[4867]: I1212 08:21:17.910691 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.159155 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329010 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329105 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329165 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dv6l\" (UniqueName: \"kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329219 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329265 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329668 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs" (OuterVolumeSpecName: "logs") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329719 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.329829 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data\") pod \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\" (UID: \"1bf77f96-18df-48d2-a7cc-7954ba9039f0\") " Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.330660 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.330685 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1bf77f96-18df-48d2-a7cc-7954ba9039f0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.334951 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts" (OuterVolumeSpecName: "scripts") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.335189 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l" (OuterVolumeSpecName: "kube-api-access-5dv6l") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "kube-api-access-5dv6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.362508 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.373211 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data" (OuterVolumeSpecName: "config-data") pod "1bf77f96-18df-48d2-a7cc-7954ba9039f0" (UID: "1bf77f96-18df-48d2-a7cc-7954ba9039f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.436050 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.436087 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.436100 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dv6l\" (UniqueName: \"kubernetes.io/projected/1bf77f96-18df-48d2-a7cc-7954ba9039f0-kube-api-access-5dv6l\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.436121 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bf77f96-18df-48d2-a7cc-7954ba9039f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.492718 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.499971 4867 generic.go:334] "Generic (PLEG): container finished" podID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerID="7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" exitCode=0 Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.500005 4867 generic.go:334] "Generic (PLEG): container finished" podID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerID="21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" exitCode=143 Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.500085 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerDied","Data":"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d"} Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.500120 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerDied","Data":"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088"} Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.500134 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1bf77f96-18df-48d2-a7cc-7954ba9039f0","Type":"ContainerDied","Data":"1c933ef4bb23ae06bc8e8c929c1890e5abe607d0458c8ae13c0b91eda6dc7f07"} Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.500152 4867 scope.go:117] "RemoveContainer" containerID="7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.501378 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.555947 4867 scope.go:117] "RemoveContainer" containerID="21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.559063 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.578042 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.588085 4867 scope.go:117] "RemoveContainer" containerID="7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" Dec 12 08:21:18 crc kubenswrapper[4867]: E1212 08:21:18.588637 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d\": container with ID starting with 7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d not found: ID does not exist" containerID="7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.588679 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d"} err="failed to get container status \"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d\": rpc error: code = NotFound desc = could not find container \"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d\": container with ID starting with 7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d not found: ID does not exist" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.588700 4867 scope.go:117] "RemoveContainer" containerID="21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" Dec 12 08:21:18 crc kubenswrapper[4867]: E1212 08:21:18.589013 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088\": container with ID starting with 21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088 not found: ID does not exist" containerID="21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.589071 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088"} err="failed to get container status \"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088\": rpc error: code = NotFound desc = could not find container \"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088\": container with ID starting with 21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088 not found: ID does not exist" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.589105 4867 scope.go:117] "RemoveContainer" containerID="7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.590712 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d"} err="failed to get container status \"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d\": rpc error: code = NotFound desc = could not find container \"7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d\": container with ID starting with 7ba56083d1335f04e74e8eaeadc0508c4943af3122a47870b48600666f195a2d not found: ID does not exist" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.590742 4867 scope.go:117] "RemoveContainer" containerID="21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.591537 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088"} err="failed to get container status \"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088\": rpc error: code = NotFound desc = could not find container \"21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088\": container with ID starting with 21646102a8d1465bdc3a407db8dfed7076ad36b9aedb976c8dadef39b6d9c088 not found: ID does not exist" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.605594 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:18 crc kubenswrapper[4867]: E1212 08:21:18.606102 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-httpd" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.606118 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-httpd" Dec 12 08:21:18 crc kubenswrapper[4867]: E1212 08:21:18.606135 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-log" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.606142 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-log" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.606410 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-httpd" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.606439 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" containerName="glance-log" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.607729 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.610833 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.612202 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.638639 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741077 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741150 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741299 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741347 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741374 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741438 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.741534 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tb2s\" (UniqueName: \"kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842770 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tb2s\" (UniqueName: \"kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842858 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842887 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842933 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842959 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.842988 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.843073 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.843879 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.843880 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.847544 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.847598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.847826 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.848751 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.849565 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf77f96-18df-48d2-a7cc-7954ba9039f0" path="/var/lib/kubelet/pods/1bf77f96-18df-48d2-a7cc-7954ba9039f0/volumes" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.850250 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e079e44-c4cb-4304-a3a6-b7814e4101d1" path="/var/lib/kubelet/pods/3e079e44-c4cb-4304-a3a6-b7814e4101d1/volumes" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.861780 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tb2s\" (UniqueName: \"kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s\") pod \"glance-default-internal-api-0\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:21:18 crc kubenswrapper[4867]: I1212 08:21:18.944728 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:19 crc kubenswrapper[4867]: I1212 08:21:19.447011 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:21:19 crc kubenswrapper[4867]: W1212 08:21:19.453869 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7852810_979e_45c7_b5f4_1f6de231699f.slice/crio-fb25bbe31618b12beb9a820c4a8da952fead6812ec870c6560a2f20710b176e9 WatchSource:0}: Error finding container fb25bbe31618b12beb9a820c4a8da952fead6812ec870c6560a2f20710b176e9: Status 404 returned error can't find the container with id fb25bbe31618b12beb9a820c4a8da952fead6812ec870c6560a2f20710b176e9 Dec 12 08:21:19 crc kubenswrapper[4867]: I1212 08:21:19.513690 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerStarted","Data":"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31"} Dec 12 08:21:19 crc kubenswrapper[4867]: I1212 08:21:19.513732 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerStarted","Data":"6e9289aab713b4b7e0d3b05711afeb4725d81a2869a020fd57f012a6dd4176e8"} Dec 12 08:21:19 crc kubenswrapper[4867]: I1212 08:21:19.516773 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerStarted","Data":"fb25bbe31618b12beb9a820c4a8da952fead6812ec870c6560a2f20710b176e9"} Dec 12 08:21:20 crc kubenswrapper[4867]: I1212 08:21:20.528563 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerStarted","Data":"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2"} Dec 12 08:21:20 crc kubenswrapper[4867]: I1212 08:21:20.530329 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerStarted","Data":"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c"} Dec 12 08:21:20 crc kubenswrapper[4867]: I1212 08:21:20.531213 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerStarted","Data":"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370"} Dec 12 08:21:20 crc kubenswrapper[4867]: I1212 08:21:20.551951 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.551931016 podStartE2EDuration="2.551931016s" podCreationTimestamp="2025-12-12 08:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:20.547301402 +0000 UTC m=+5568.118682691" watchObservedRunningTime="2025-12-12 08:21:20.551931016 +0000 UTC m=+5568.123312285" Dec 12 08:21:20 crc kubenswrapper[4867]: I1212 08:21:20.578046 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.5777822969999997 podStartE2EDuration="3.577782297s" podCreationTimestamp="2025-12-12 08:21:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:20.573535322 +0000 UTC m=+5568.144916591" watchObservedRunningTime="2025-12-12 08:21:20.577782297 +0000 UTC m=+5568.149163566" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.079388 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.153275 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.153520 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="dnsmasq-dns" containerID="cri-o://73cd0d4fc21d8b40133a343b64467b830c47f69e8b79463d7de44050ee16cf0c" gracePeriod=10 Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.565777 4867 generic.go:334] "Generic (PLEG): container finished" podID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerID="73cd0d4fc21d8b40133a343b64467b830c47f69e8b79463d7de44050ee16cf0c" exitCode=0 Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.565877 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" event={"ID":"d741ed2b-868d-4a97-bde9-aa5f548f9e45","Type":"ContainerDied","Data":"73cd0d4fc21d8b40133a343b64467b830c47f69e8b79463d7de44050ee16cf0c"} Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.566123 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" event={"ID":"d741ed2b-868d-4a97-bde9-aa5f548f9e45","Type":"ContainerDied","Data":"477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae"} Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.566187 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="477105695eb5e6307f71d78963221928b0c0fa2da9f7caaa420aebfde87360ae" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.646734 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.743696 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb\") pod \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.743740 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb\") pod \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.743758 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc\") pod \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.743856 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q2qt\" (UniqueName: \"kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt\") pod \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.743921 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config\") pod \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\" (UID: \"d741ed2b-868d-4a97-bde9-aa5f548f9e45\") " Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.749673 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt" (OuterVolumeSpecName: "kube-api-access-8q2qt") pod "d741ed2b-868d-4a97-bde9-aa5f548f9e45" (UID: "d741ed2b-868d-4a97-bde9-aa5f548f9e45"). InnerVolumeSpecName "kube-api-access-8q2qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.791161 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d741ed2b-868d-4a97-bde9-aa5f548f9e45" (UID: "d741ed2b-868d-4a97-bde9-aa5f548f9e45"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.793420 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d741ed2b-868d-4a97-bde9-aa5f548f9e45" (UID: "d741ed2b-868d-4a97-bde9-aa5f548f9e45"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.794985 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config" (OuterVolumeSpecName: "config") pod "d741ed2b-868d-4a97-bde9-aa5f548f9e45" (UID: "d741ed2b-868d-4a97-bde9-aa5f548f9e45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.806270 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d741ed2b-868d-4a97-bde9-aa5f548f9e45" (UID: "d741ed2b-868d-4a97-bde9-aa5f548f9e45"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.845511 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.845558 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.845571 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.845584 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q2qt\" (UniqueName: \"kubernetes.io/projected/d741ed2b-868d-4a97-bde9-aa5f548f9e45-kube-api-access-8q2qt\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:24 crc kubenswrapper[4867]: I1212 08:21:24.845596 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d741ed2b-868d-4a97-bde9-aa5f548f9e45-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:25 crc kubenswrapper[4867]: I1212 08:21:25.573828 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5884c65cf-k9bnc" Dec 12 08:21:25 crc kubenswrapper[4867]: I1212 08:21:25.597730 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:21:25 crc kubenswrapper[4867]: I1212 08:21:25.605216 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5884c65cf-k9bnc"] Dec 12 08:21:26 crc kubenswrapper[4867]: I1212 08:21:26.838815 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:21:26 crc kubenswrapper[4867]: E1212 08:21:26.839085 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:21:26 crc kubenswrapper[4867]: I1212 08:21:26.849657 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" path="/var/lib/kubelet/pods/d741ed2b-868d-4a97-bde9-aa5f548f9e45/volumes" Dec 12 08:21:27 crc kubenswrapper[4867]: I1212 08:21:27.910969 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 08:21:27 crc kubenswrapper[4867]: I1212 08:21:27.911327 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 08:21:27 crc kubenswrapper[4867]: I1212 08:21:27.942732 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 08:21:27 crc kubenswrapper[4867]: I1212 08:21:27.950209 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.604401 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.604483 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.945037 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.945106 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.979364 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:28 crc kubenswrapper[4867]: I1212 08:21:28.992585 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:29 crc kubenswrapper[4867]: I1212 08:21:29.613924 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:29 crc kubenswrapper[4867]: I1212 08:21:29.613973 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:30 crc kubenswrapper[4867]: I1212 08:21:30.853114 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 08:21:30 crc kubenswrapper[4867]: I1212 08:21:30.853214 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 08:21:32 crc kubenswrapper[4867]: I1212 08:21:32.451599 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:32 crc kubenswrapper[4867]: I1212 08:21:32.452351 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 08:21:32 crc kubenswrapper[4867]: I1212 08:21:32.454068 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.503270 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-rbk8h"] Dec 12 08:21:38 crc kubenswrapper[4867]: E1212 08:21:38.504089 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="dnsmasq-dns" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.504101 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="dnsmasq-dns" Dec 12 08:21:38 crc kubenswrapper[4867]: E1212 08:21:38.504112 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="init" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.504118 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="init" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.504369 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d741ed2b-868d-4a97-bde9-aa5f548f9e45" containerName="dnsmasq-dns" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.504982 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.512173 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rbk8h"] Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.601608 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f55b-account-create-update-db6tz"] Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.603186 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.606298 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.606405 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9tj\" (UniqueName: \"kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.609196 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.610443 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f55b-account-create-update-db6tz"] Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.707913 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn9zc\" (UniqueName: \"kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.707976 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.708010 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.708126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9tj\" (UniqueName: \"kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.709000 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.729051 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9tj\" (UniqueName: \"kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj\") pod \"placement-db-create-rbk8h\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.810063 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn9zc\" (UniqueName: \"kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.810115 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.810908 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.824001 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.836685 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn9zc\" (UniqueName: \"kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc\") pod \"placement-f55b-account-create-update-db6tz\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.838246 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:21:38 crc kubenswrapper[4867]: I1212 08:21:38.924644 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.301723 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-rbk8h"] Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.403680 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f55b-account-create-update-db6tz"] Dec 12 08:21:39 crc kubenswrapper[4867]: W1212 08:21:39.409194 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a1ef07e_6698_40ab_9251_ca02f2e2726f.slice/crio-6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d WatchSource:0}: Error finding container 6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d: Status 404 returned error can't find the container with id 6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.719327 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f55b-account-create-update-db6tz" event={"ID":"9a1ef07e-6698-40ab-9251-ca02f2e2726f","Type":"ContainerStarted","Data":"6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d"} Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.722967 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52"} Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.725973 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rbk8h" event={"ID":"f97a1ec6-d7fc-4f2d-a063-891aee80a17b","Type":"ContainerStarted","Data":"41cd9fb6589c4a22a975a78cd6c202ebd083af10a78d50917a3bc11e69b1b1d0"} Dec 12 08:21:39 crc kubenswrapper[4867]: I1212 08:21:39.726027 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rbk8h" event={"ID":"f97a1ec6-d7fc-4f2d-a063-891aee80a17b","Type":"ContainerStarted","Data":"75bf5966108e4d0126ab1bb051997f387589897850460cd051d1a295fe0eef74"} Dec 12 08:21:40 crc kubenswrapper[4867]: I1212 08:21:40.758378 4867 generic.go:334] "Generic (PLEG): container finished" podID="f97a1ec6-d7fc-4f2d-a063-891aee80a17b" containerID="41cd9fb6589c4a22a975a78cd6c202ebd083af10a78d50917a3bc11e69b1b1d0" exitCode=0 Dec 12 08:21:40 crc kubenswrapper[4867]: I1212 08:21:40.758528 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rbk8h" event={"ID":"f97a1ec6-d7fc-4f2d-a063-891aee80a17b","Type":"ContainerDied","Data":"41cd9fb6589c4a22a975a78cd6c202ebd083af10a78d50917a3bc11e69b1b1d0"} Dec 12 08:21:40 crc kubenswrapper[4867]: I1212 08:21:40.766089 4867 generic.go:334] "Generic (PLEG): container finished" podID="9a1ef07e-6698-40ab-9251-ca02f2e2726f" containerID="ffdd0a96d087e818753d60df3de929e64e6477dcef0a15cb4b5d224a5762569c" exitCode=0 Dec 12 08:21:40 crc kubenswrapper[4867]: I1212 08:21:40.766158 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f55b-account-create-update-db6tz" event={"ID":"9a1ef07e-6698-40ab-9251-ca02f2e2726f","Type":"ContainerDied","Data":"ffdd0a96d087e818753d60df3de929e64e6477dcef0a15cb4b5d224a5762569c"} Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.159126 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.170306 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.284565 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts\") pod \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.284616 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn9zc\" (UniqueName: \"kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc\") pod \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.284652 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts\") pod \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\" (UID: \"9a1ef07e-6698-40ab-9251-ca02f2e2726f\") " Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.284715 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg9tj\" (UniqueName: \"kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj\") pod \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\" (UID: \"f97a1ec6-d7fc-4f2d-a063-891aee80a17b\") " Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.285599 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f97a1ec6-d7fc-4f2d-a063-891aee80a17b" (UID: "f97a1ec6-d7fc-4f2d-a063-891aee80a17b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.285821 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a1ef07e-6698-40ab-9251-ca02f2e2726f" (UID: "9a1ef07e-6698-40ab-9251-ca02f2e2726f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.291522 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj" (OuterVolumeSpecName: "kube-api-access-wg9tj") pod "f97a1ec6-d7fc-4f2d-a063-891aee80a17b" (UID: "f97a1ec6-d7fc-4f2d-a063-891aee80a17b"). InnerVolumeSpecName "kube-api-access-wg9tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.291588 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc" (OuterVolumeSpecName: "kube-api-access-gn9zc") pod "9a1ef07e-6698-40ab-9251-ca02f2e2726f" (UID: "9a1ef07e-6698-40ab-9251-ca02f2e2726f"). InnerVolumeSpecName "kube-api-access-gn9zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.387418 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.387457 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn9zc\" (UniqueName: \"kubernetes.io/projected/9a1ef07e-6698-40ab-9251-ca02f2e2726f-kube-api-access-gn9zc\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.388332 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1ef07e-6698-40ab-9251-ca02f2e2726f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.388349 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg9tj\" (UniqueName: \"kubernetes.io/projected/f97a1ec6-d7fc-4f2d-a063-891aee80a17b-kube-api-access-wg9tj\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.786214 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-rbk8h" event={"ID":"f97a1ec6-d7fc-4f2d-a063-891aee80a17b","Type":"ContainerDied","Data":"75bf5966108e4d0126ab1bb051997f387589897850460cd051d1a295fe0eef74"} Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.786503 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75bf5966108e4d0126ab1bb051997f387589897850460cd051d1a295fe0eef74" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.786256 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-rbk8h" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.788297 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f55b-account-create-update-db6tz" Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.788296 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f55b-account-create-update-db6tz" event={"ID":"9a1ef07e-6698-40ab-9251-ca02f2e2726f","Type":"ContainerDied","Data":"6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d"} Dec 12 08:21:42 crc kubenswrapper[4867]: I1212 08:21:42.788422 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6523ce996fb1803a28c94a45bac32275080dae7b4dd3ef9acea8f6fb93f3829d" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.837125 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:21:43 crc kubenswrapper[4867]: E1212 08:21:43.838082 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1ef07e-6698-40ab-9251-ca02f2e2726f" containerName="mariadb-account-create-update" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.838100 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1ef07e-6698-40ab-9251-ca02f2e2726f" containerName="mariadb-account-create-update" Dec 12 08:21:43 crc kubenswrapper[4867]: E1212 08:21:43.838135 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97a1ec6-d7fc-4f2d-a063-891aee80a17b" containerName="mariadb-database-create" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.838144 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97a1ec6-d7fc-4f2d-a063-891aee80a17b" containerName="mariadb-database-create" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.838374 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97a1ec6-d7fc-4f2d-a063-891aee80a17b" containerName="mariadb-database-create" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.838408 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1ef07e-6698-40ab-9251-ca02f2e2726f" containerName="mariadb-account-create-update" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.839865 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.860328 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.878163 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zzx5d"] Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.879393 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.883754 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.884165 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xmvv7" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.884341 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.894428 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zzx5d"] Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.925424 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwbln\" (UniqueName: \"kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.925557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.925608 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.925731 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:43 crc kubenswrapper[4867]: I1212 08:21:43.925788 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028095 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028180 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028217 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028262 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028316 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028340 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028370 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcfbl\" (UniqueName: \"kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028402 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028443 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.028689 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwbln\" (UniqueName: \"kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.029310 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.029398 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.029442 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.030593 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.046881 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwbln\" (UniqueName: \"kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln\") pod \"dnsmasq-dns-95c76c545-ttnbc\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.130346 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.130422 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.131033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.131112 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.131533 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcfbl\" (UniqueName: \"kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.131602 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.134083 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.134267 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.134304 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.146903 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcfbl\" (UniqueName: \"kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl\") pod \"placement-db-sync-zzx5d\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.179410 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.207261 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.714513 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.764343 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zzx5d"] Dec 12 08:21:44 crc kubenswrapper[4867]: W1212 08:21:44.766003 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62e38feb_db94_479d_9856_cd92df3741e1.slice/crio-f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b WatchSource:0}: Error finding container f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b: Status 404 returned error can't find the container with id f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.809792 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zzx5d" event={"ID":"62e38feb-db94-479d-9856-cd92df3741e1","Type":"ContainerStarted","Data":"f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b"} Dec 12 08:21:44 crc kubenswrapper[4867]: I1212 08:21:44.812051 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" event={"ID":"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979","Type":"ContainerStarted","Data":"cf01d8d82a00d58f7006e34f961af28ad3931872ede85edd7e38059b05b6cd0f"} Dec 12 08:21:45 crc kubenswrapper[4867]: I1212 08:21:45.823012 4867 generic.go:334] "Generic (PLEG): container finished" podID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerID="663ce92697c58fd77df4f91db759ad92c6f4bd238864646cba8f2221813208a5" exitCode=0 Dec 12 08:21:45 crc kubenswrapper[4867]: I1212 08:21:45.823161 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" event={"ID":"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979","Type":"ContainerDied","Data":"663ce92697c58fd77df4f91db759ad92c6f4bd238864646cba8f2221813208a5"} Dec 12 08:21:46 crc kubenswrapper[4867]: I1212 08:21:46.836269 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" event={"ID":"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979","Type":"ContainerStarted","Data":"dfc462fba9233854bf0c33db5ad6610b989987dcdb0f2333591adfebbc9b788d"} Dec 12 08:21:46 crc kubenswrapper[4867]: I1212 08:21:46.836658 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:48 crc kubenswrapper[4867]: I1212 08:21:48.852241 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zzx5d" event={"ID":"62e38feb-db94-479d-9856-cd92df3741e1","Type":"ContainerStarted","Data":"3582b78b6abced80afd095968cb3ce17490aefa756cb1c6f43443f966674e197"} Dec 12 08:21:48 crc kubenswrapper[4867]: I1212 08:21:48.876625 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" podStartSLOduration=5.87660836 podStartE2EDuration="5.87660836s" podCreationTimestamp="2025-12-12 08:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:46.860657689 +0000 UTC m=+5594.432038978" watchObservedRunningTime="2025-12-12 08:21:48.87660836 +0000 UTC m=+5596.447989629" Dec 12 08:21:50 crc kubenswrapper[4867]: I1212 08:21:50.869825 4867 generic.go:334] "Generic (PLEG): container finished" podID="62e38feb-db94-479d-9856-cd92df3741e1" containerID="3582b78b6abced80afd095968cb3ce17490aefa756cb1c6f43443f966674e197" exitCode=0 Dec 12 08:21:50 crc kubenswrapper[4867]: I1212 08:21:50.869910 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zzx5d" event={"ID":"62e38feb-db94-479d-9856-cd92df3741e1","Type":"ContainerDied","Data":"3582b78b6abced80afd095968cb3ce17490aefa756cb1c6f43443f966674e197"} Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.265261 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.290525 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs\") pod \"62e38feb-db94-479d-9856-cd92df3741e1\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.290602 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle\") pod \"62e38feb-db94-479d-9856-cd92df3741e1\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.290835 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts\") pod \"62e38feb-db94-479d-9856-cd92df3741e1\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.290899 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcfbl\" (UniqueName: \"kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl\") pod \"62e38feb-db94-479d-9856-cd92df3741e1\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.290927 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data\") pod \"62e38feb-db94-479d-9856-cd92df3741e1\" (UID: \"62e38feb-db94-479d-9856-cd92df3741e1\") " Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.291066 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs" (OuterVolumeSpecName: "logs") pod "62e38feb-db94-479d-9856-cd92df3741e1" (UID: "62e38feb-db94-479d-9856-cd92df3741e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.291810 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62e38feb-db94-479d-9856-cd92df3741e1-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.303809 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl" (OuterVolumeSpecName: "kube-api-access-xcfbl") pod "62e38feb-db94-479d-9856-cd92df3741e1" (UID: "62e38feb-db94-479d-9856-cd92df3741e1"). InnerVolumeSpecName "kube-api-access-xcfbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.324630 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62e38feb-db94-479d-9856-cd92df3741e1" (UID: "62e38feb-db94-479d-9856-cd92df3741e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.329514 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts" (OuterVolumeSpecName: "scripts") pod "62e38feb-db94-479d-9856-cd92df3741e1" (UID: "62e38feb-db94-479d-9856-cd92df3741e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.332692 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data" (OuterVolumeSpecName: "config-data") pod "62e38feb-db94-479d-9856-cd92df3741e1" (UID: "62e38feb-db94-479d-9856-cd92df3741e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.393479 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.393943 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.393998 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcfbl\" (UniqueName: \"kubernetes.io/projected/62e38feb-db94-479d-9856-cd92df3741e1-kube-api-access-xcfbl\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.394049 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62e38feb-db94-479d-9856-cd92df3741e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.909384 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zzx5d" event={"ID":"62e38feb-db94-479d-9856-cd92df3741e1","Type":"ContainerDied","Data":"f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b"} Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.909424 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2d9ec0f1cbd2433e3c9ea5657e905ceb1868a52d02995682f32f9a0acc92f3b" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.909482 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zzx5d" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.974532 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68c67cf9f8-5kb5p"] Dec 12 08:21:52 crc kubenswrapper[4867]: E1212 08:21:52.975034 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e38feb-db94-479d-9856-cd92df3741e1" containerName="placement-db-sync" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.975059 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e38feb-db94-479d-9856-cd92df3741e1" containerName="placement-db-sync" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.975290 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e38feb-db94-479d-9856-cd92df3741e1" containerName="placement-db-sync" Dec 12 08:21:52 crc kubenswrapper[4867]: I1212 08:21:52.976529 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:52.981673 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:52.981677 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:52.981950 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xmvv7" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:52.982058 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:52.983978 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.024568 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-public-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.024767 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjpk\" (UniqueName: \"kubernetes.io/projected/f0286224-fef5-43fc-895c-15806d60a789-kube-api-access-zxjpk\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.024951 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0286224-fef5-43fc-895c-15806d60a789-logs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.025009 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-scripts\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.025065 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-config-data\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.025219 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-combined-ca-bundle\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.025297 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-internal-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.030470 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68c67cf9f8-5kb5p"] Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.127908 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-public-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128107 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjpk\" (UniqueName: \"kubernetes.io/projected/f0286224-fef5-43fc-895c-15806d60a789-kube-api-access-zxjpk\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128193 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0286224-fef5-43fc-895c-15806d60a789-logs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128246 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-scripts\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128275 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-config-data\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128346 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-combined-ca-bundle\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.128416 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-internal-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.129273 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0286224-fef5-43fc-895c-15806d60a789-logs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.132772 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-public-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.132817 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-scripts\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.134036 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-combined-ca-bundle\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.134870 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-config-data\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.139071 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0286224-fef5-43fc-895c-15806d60a789-internal-tls-certs\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.145675 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjpk\" (UniqueName: \"kubernetes.io/projected/f0286224-fef5-43fc-895c-15806d60a789-kube-api-access-zxjpk\") pod \"placement-68c67cf9f8-5kb5p\" (UID: \"f0286224-fef5-43fc-895c-15806d60a789\") " pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.342534 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.797435 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68c67cf9f8-5kb5p"] Dec 12 08:21:53 crc kubenswrapper[4867]: I1212 08:21:53.922705 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68c67cf9f8-5kb5p" event={"ID":"f0286224-fef5-43fc-895c-15806d60a789","Type":"ContainerStarted","Data":"706546d6fa32d37025cc8e47cedeed5dc2536b6f6c16da5ba6b36bc616d31af0"} Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.181237 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.250650 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.258169 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="dnsmasq-dns" containerID="cri-o://5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862" gracePeriod=10 Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.750991 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.875152 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc\") pod \"106d309a-9808-4be0-a2ee-6a7522ebc357\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.875220 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb\") pod \"106d309a-9808-4be0-a2ee-6a7522ebc357\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.875303 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74j9g\" (UniqueName: \"kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g\") pod \"106d309a-9808-4be0-a2ee-6a7522ebc357\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.875341 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config\") pod \"106d309a-9808-4be0-a2ee-6a7522ebc357\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.875459 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb\") pod \"106d309a-9808-4be0-a2ee-6a7522ebc357\" (UID: \"106d309a-9808-4be0-a2ee-6a7522ebc357\") " Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.882169 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g" (OuterVolumeSpecName: "kube-api-access-74j9g") pod "106d309a-9808-4be0-a2ee-6a7522ebc357" (UID: "106d309a-9808-4be0-a2ee-6a7522ebc357"). InnerVolumeSpecName "kube-api-access-74j9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.926721 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "106d309a-9808-4be0-a2ee-6a7522ebc357" (UID: "106d309a-9808-4be0-a2ee-6a7522ebc357"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.933613 4867 generic.go:334] "Generic (PLEG): container finished" podID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerID="5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862" exitCode=0 Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.933677 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.933678 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" event={"ID":"106d309a-9808-4be0-a2ee-6a7522ebc357","Type":"ContainerDied","Data":"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862"} Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.933778 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f9f666b79-bxtzx" event={"ID":"106d309a-9808-4be0-a2ee-6a7522ebc357","Type":"ContainerDied","Data":"349d3cf77a9d14150a34090dabf5fe96318317af5edeb159963bd4bfe3c08339"} Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.933798 4867 scope.go:117] "RemoveContainer" containerID="5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.935911 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68c67cf9f8-5kb5p" event={"ID":"f0286224-fef5-43fc-895c-15806d60a789","Type":"ContainerStarted","Data":"6a19a78c5ecdbc8639bcc7e9de1e31861ee4b4887e9e585b715e459d323d06cd"} Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.935938 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68c67cf9f8-5kb5p" event={"ID":"f0286224-fef5-43fc-895c-15806d60a789","Type":"ContainerStarted","Data":"76211fa539db719fbc091cd956a9d41b8ab7bdf4a8d62fc2738a36c033e69e95"} Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.936090 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.936107 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.942725 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config" (OuterVolumeSpecName: "config") pod "106d309a-9808-4be0-a2ee-6a7522ebc357" (UID: "106d309a-9808-4be0-a2ee-6a7522ebc357"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.943070 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "106d309a-9808-4be0-a2ee-6a7522ebc357" (UID: "106d309a-9808-4be0-a2ee-6a7522ebc357"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.947957 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "106d309a-9808-4be0-a2ee-6a7522ebc357" (UID: "106d309a-9808-4be0-a2ee-6a7522ebc357"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.969207 4867 scope.go:117] "RemoveContainer" containerID="7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.969959 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68c67cf9f8-5kb5p" podStartSLOduration=2.969946322 podStartE2EDuration="2.969946322s" podCreationTimestamp="2025-12-12 08:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:21:54.95978121 +0000 UTC m=+5602.531162489" watchObservedRunningTime="2025-12-12 08:21:54.969946322 +0000 UTC m=+5602.541327591" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.981427 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.981463 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.981475 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.981488 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74j9g\" (UniqueName: \"kubernetes.io/projected/106d309a-9808-4be0-a2ee-6a7522ebc357-kube-api-access-74j9g\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:54 crc kubenswrapper[4867]: I1212 08:21:54.981500 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/106d309a-9808-4be0-a2ee-6a7522ebc357-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.007699 4867 scope.go:117] "RemoveContainer" containerID="5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862" Dec 12 08:21:55 crc kubenswrapper[4867]: E1212 08:21:55.008586 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862\": container with ID starting with 5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862 not found: ID does not exist" containerID="5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862" Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.008772 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862"} err="failed to get container status \"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862\": rpc error: code = NotFound desc = could not find container \"5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862\": container with ID starting with 5224427fd59a560306042e45c12193e54473e7a03d6aca3ca2594109c8b54862 not found: ID does not exist" Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.008799 4867 scope.go:117] "RemoveContainer" containerID="7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278" Dec 12 08:21:55 crc kubenswrapper[4867]: E1212 08:21:55.009520 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278\": container with ID starting with 7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278 not found: ID does not exist" containerID="7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278" Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.009548 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278"} err="failed to get container status \"7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278\": rpc error: code = NotFound desc = could not find container \"7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278\": container with ID starting with 7606ef8fe224f36f85f328f176be5cf2ac4a59831a1baeea1570fae352861278 not found: ID does not exist" Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.277261 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:55 crc kubenswrapper[4867]: I1212 08:21:55.286143 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f9f666b79-bxtzx"] Dec 12 08:21:56 crc kubenswrapper[4867]: I1212 08:21:56.848684 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" path="/var/lib/kubelet/pods/106d309a-9808-4be0-a2ee-6a7522ebc357/volumes" Dec 12 08:22:24 crc kubenswrapper[4867]: I1212 08:22:24.420011 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:22:25 crc kubenswrapper[4867]: I1212 08:22:25.499061 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68c67cf9f8-5kb5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.449724 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-skvtb"] Dec 12 08:22:46 crc kubenswrapper[4867]: E1212 08:22:46.450613 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="init" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.450625 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="init" Dec 12 08:22:46 crc kubenswrapper[4867]: E1212 08:22:46.450650 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="dnsmasq-dns" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.450656 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="dnsmasq-dns" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.450827 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="106d309a-9808-4be0-a2ee-6a7522ebc357" containerName="dnsmasq-dns" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.451521 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.460993 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-skvtb"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.550577 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5657-account-create-update-tl2gr"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.552153 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.553457 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqt6g\" (UniqueName: \"kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.553693 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.553836 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.559142 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-r6b5p"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.560968 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.571646 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r6b5p"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.581491 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5657-account-create-update-tl2gr"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.655758 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.655952 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.655993 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-264zg\" (UniqueName: \"kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.656037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqt6g\" (UniqueName: \"kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.656626 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.667016 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-sgljh"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.668249 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.676785 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sgljh"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.683027 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqt6g\" (UniqueName: \"kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g\") pod \"nova-api-db-create-skvtb\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.755183 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0e7e-account-create-update-g6wnx"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.757153 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.757999 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.758199 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.758256 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-264zg\" (UniqueName: \"kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.758287 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q94db\" (UniqueName: \"kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.759033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.760319 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.772192 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0e7e-account-create-update-g6wnx"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.784493 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.795038 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-264zg\" (UniqueName: \"kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg\") pod \"nova-api-5657-account-create-update-tl2gr\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.861020 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.861718 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q94db\" (UniqueName: \"kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.861797 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.861826 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.861870 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkbt8\" (UniqueName: \"kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.862026 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc65s\" (UniqueName: \"kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.862764 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.875821 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.885284 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q94db\" (UniqueName: \"kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db\") pod \"nova-cell0-db-create-r6b5p\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.963625 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc65s\" (UniqueName: \"kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.963745 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.963852 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.963907 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkbt8\" (UniqueName: \"kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.964692 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.964824 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.967973 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8bcb-account-create-update-cf9b2"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.969817 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.975389 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.987061 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bcb-account-create-update-cf9b2"] Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.992556 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc65s\" (UniqueName: \"kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s\") pod \"nova-cell1-db-create-sgljh\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:46 crc kubenswrapper[4867]: I1212 08:22:46.995138 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkbt8\" (UniqueName: \"kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8\") pod \"nova-cell0-0e7e-account-create-update-g6wnx\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.051107 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.078799 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.169446 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zckc\" (UniqueName: \"kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.169557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.181808 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.271740 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zckc\" (UniqueName: \"kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.271826 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.273056 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.295047 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zckc\" (UniqueName: \"kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc\") pod \"nova-cell1-8bcb-account-create-update-cf9b2\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.297094 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-skvtb"] Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.314501 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.387465 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-skvtb" event={"ID":"c59d7f39-1709-49ba-8623-41f86f89e219","Type":"ContainerStarted","Data":"3873bb99beab6b3227e9593e3c15492efe8b0d9ed8e8593855048ba643c45b04"} Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.442664 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5657-account-create-update-tl2gr"] Dec 12 08:22:47 crc kubenswrapper[4867]: W1212 08:22:47.451903 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab3d2425_2baf_4842_ac92_256766803f72.slice/crio-db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089 WatchSource:0}: Error finding container db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089: Status 404 returned error can't find the container with id db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089 Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.576670 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0e7e-account-create-update-g6wnx"] Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.647941 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8bcb-account-create-update-cf9b2"] Dec 12 08:22:47 crc kubenswrapper[4867]: W1212 08:22:47.663432 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd48bc89d_ef25_4b10_96f0_be3a93c6eff1.slice/crio-9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca WatchSource:0}: Error finding container 9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca: Status 404 returned error can't find the container with id 9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.669896 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-sgljh"] Dec 12 08:22:47 crc kubenswrapper[4867]: W1212 08:22:47.676462 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod565d5803_c324_4497_a928_a84f18aaea1b.slice/crio-bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097 WatchSource:0}: Error finding container bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097: Status 404 returned error can't find the container with id bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097 Dec 12 08:22:47 crc kubenswrapper[4867]: I1212 08:22:47.740607 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-r6b5p"] Dec 12 08:22:47 crc kubenswrapper[4867]: W1212 08:22:47.744144 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa6cd035_6852_4418_ac56_33fa1f5c2a45.slice/crio-084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c WatchSource:0}: Error finding container 084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c: Status 404 returned error can't find the container with id 084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c Dec 12 08:22:47 crc kubenswrapper[4867]: E1212 08:22:47.959910 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc59d7f39_1709_49ba_8623_41f86f89e219.slice/crio-conmon-2279e76358e2439e53b3b8f373b48ec98fedac9fb7af5fc9a20c1a46bfded688.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab3d2425_2baf_4842_ac92_256766803f72.slice/crio-conmon-0d47260215dbab76368bdcd4eafcd308c2f1399f66909dce69490a6928865fdd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc59d7f39_1709_49ba_8623_41f86f89e219.slice/crio-2279e76358e2439e53b3b8f373b48ec98fedac9fb7af5fc9a20c1a46bfded688.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.401373 4867 generic.go:334] "Generic (PLEG): container finished" podID="d48bc89d-ef25-4b10-96f0-be3a93c6eff1" containerID="8e7b7f40702c400378f429ccaed298bb3531c5cb8102bf9b995945297ce0904b" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.401465 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" event={"ID":"d48bc89d-ef25-4b10-96f0-be3a93c6eff1","Type":"ContainerDied","Data":"8e7b7f40702c400378f429ccaed298bb3531c5cb8102bf9b995945297ce0904b"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.401522 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" event={"ID":"d48bc89d-ef25-4b10-96f0-be3a93c6eff1","Type":"ContainerStarted","Data":"9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.403950 4867 generic.go:334] "Generic (PLEG): container finished" podID="565d5803-c324-4497-a928-a84f18aaea1b" containerID="0d81a53e3e4ed7888e30655ed6df26c22647349f6dbf6483af939613c7b1782c" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.404031 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sgljh" event={"ID":"565d5803-c324-4497-a928-a84f18aaea1b","Type":"ContainerDied","Data":"0d81a53e3e4ed7888e30655ed6df26c22647349f6dbf6483af939613c7b1782c"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.404053 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sgljh" event={"ID":"565d5803-c324-4497-a928-a84f18aaea1b","Type":"ContainerStarted","Data":"bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.407077 4867 generic.go:334] "Generic (PLEG): container finished" podID="fa6cd035-6852-4418-ac56-33fa1f5c2a45" containerID="4b6aff1f47fdb4eef73e4e956bac34768aa59938f10924fccea1167da4751cb5" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.407121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r6b5p" event={"ID":"fa6cd035-6852-4418-ac56-33fa1f5c2a45","Type":"ContainerDied","Data":"4b6aff1f47fdb4eef73e4e956bac34768aa59938f10924fccea1167da4751cb5"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.407143 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r6b5p" event={"ID":"fa6cd035-6852-4418-ac56-33fa1f5c2a45","Type":"ContainerStarted","Data":"084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.408873 4867 generic.go:334] "Generic (PLEG): container finished" podID="c59d7f39-1709-49ba-8623-41f86f89e219" containerID="2279e76358e2439e53b3b8f373b48ec98fedac9fb7af5fc9a20c1a46bfded688" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.408911 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-skvtb" event={"ID":"c59d7f39-1709-49ba-8623-41f86f89e219","Type":"ContainerDied","Data":"2279e76358e2439e53b3b8f373b48ec98fedac9fb7af5fc9a20c1a46bfded688"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.410853 4867 generic.go:334] "Generic (PLEG): container finished" podID="a7f689bc-82d9-4867-8505-8ccc6eb237ce" containerID="128c9e5af0db226180955e86c009c965e29b78ae1b0a222dd70c5084c5e58fb1" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.410896 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" event={"ID":"a7f689bc-82d9-4867-8505-8ccc6eb237ce","Type":"ContainerDied","Data":"128c9e5af0db226180955e86c009c965e29b78ae1b0a222dd70c5084c5e58fb1"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.410917 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" event={"ID":"a7f689bc-82d9-4867-8505-8ccc6eb237ce","Type":"ContainerStarted","Data":"42a7a9b900f3e2c5b1a453f133e79a25e6b73e0c481b8aff7e4bfd7771f3ffb2"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.415566 4867 generic.go:334] "Generic (PLEG): container finished" podID="ab3d2425-2baf-4842-ac92-256766803f72" containerID="0d47260215dbab76368bdcd4eafcd308c2f1399f66909dce69490a6928865fdd" exitCode=0 Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.415614 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5657-account-create-update-tl2gr" event={"ID":"ab3d2425-2baf-4842-ac92-256766803f72","Type":"ContainerDied","Data":"0d47260215dbab76368bdcd4eafcd308c2f1399f66909dce69490a6928865fdd"} Dec 12 08:22:48 crc kubenswrapper[4867]: I1212 08:22:48.415639 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5657-account-create-update-tl2gr" event={"ID":"ab3d2425-2baf-4842-ac92-256766803f72","Type":"ContainerStarted","Data":"db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089"} Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.810122 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.917796 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts\") pod \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.917956 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q94db\" (UniqueName: \"kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db\") pod \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\" (UID: \"fa6cd035-6852-4418-ac56-33fa1f5c2a45\") " Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.918868 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa6cd035-6852-4418-ac56-33fa1f5c2a45" (UID: "fa6cd035-6852-4418-ac56-33fa1f5c2a45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.925203 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db" (OuterVolumeSpecName: "kube-api-access-q94db") pod "fa6cd035-6852-4418-ac56-33fa1f5c2a45" (UID: "fa6cd035-6852-4418-ac56-33fa1f5c2a45"). InnerVolumeSpecName "kube-api-access-q94db". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.963924 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.973015 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.981321 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:49 crc kubenswrapper[4867]: I1212 08:22:49.996059 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.007276 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.020528 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa6cd035-6852-4418-ac56-33fa1f5c2a45-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.020560 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q94db\" (UniqueName: \"kubernetes.io/projected/fa6cd035-6852-4418-ac56-33fa1f5c2a45-kube-api-access-q94db\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.121479 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zckc\" (UniqueName: \"kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc\") pod \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.121765 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkbt8\" (UniqueName: \"kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8\") pod \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.121873 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts\") pod \"565d5803-c324-4497-a928-a84f18aaea1b\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.121987 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqt6g\" (UniqueName: \"kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g\") pod \"c59d7f39-1709-49ba-8623-41f86f89e219\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122083 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc65s\" (UniqueName: \"kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s\") pod \"565d5803-c324-4497-a928-a84f18aaea1b\" (UID: \"565d5803-c324-4497-a928-a84f18aaea1b\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122262 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts\") pod \"c59d7f39-1709-49ba-8623-41f86f89e219\" (UID: \"c59d7f39-1709-49ba-8623-41f86f89e219\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122390 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts\") pod \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\" (UID: \"a7f689bc-82d9-4867-8505-8ccc6eb237ce\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122484 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts\") pod \"ab3d2425-2baf-4842-ac92-256766803f72\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122584 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts\") pod \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\" (UID: \"d48bc89d-ef25-4b10-96f0-be3a93c6eff1\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122666 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c59d7f39-1709-49ba-8623-41f86f89e219" (UID: "c59d7f39-1709-49ba-8623-41f86f89e219"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122707 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "565d5803-c324-4497-a928-a84f18aaea1b" (UID: "565d5803-c324-4497-a928-a84f18aaea1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.122830 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-264zg\" (UniqueName: \"kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg\") pod \"ab3d2425-2baf-4842-ac92-256766803f72\" (UID: \"ab3d2425-2baf-4842-ac92-256766803f72\") " Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123037 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ab3d2425-2baf-4842-ac92-256766803f72" (UID: "ab3d2425-2baf-4842-ac92-256766803f72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123452 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d48bc89d-ef25-4b10-96f0-be3a93c6eff1" (UID: "d48bc89d-ef25-4b10-96f0-be3a93c6eff1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123462 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7f689bc-82d9-4867-8505-8ccc6eb237ce" (UID: "a7f689bc-82d9-4867-8505-8ccc6eb237ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123634 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/565d5803-c324-4497-a928-a84f18aaea1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123725 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c59d7f39-1709-49ba-8623-41f86f89e219-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.123953 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab3d2425-2baf-4842-ac92-256766803f72-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.124753 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc" (OuterVolumeSpecName: "kube-api-access-7zckc") pod "d48bc89d-ef25-4b10-96f0-be3a93c6eff1" (UID: "d48bc89d-ef25-4b10-96f0-be3a93c6eff1"). InnerVolumeSpecName "kube-api-access-7zckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.124868 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8" (OuterVolumeSpecName: "kube-api-access-gkbt8") pod "a7f689bc-82d9-4867-8505-8ccc6eb237ce" (UID: "a7f689bc-82d9-4867-8505-8ccc6eb237ce"). InnerVolumeSpecName "kube-api-access-gkbt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.125174 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g" (OuterVolumeSpecName: "kube-api-access-jqt6g") pod "c59d7f39-1709-49ba-8623-41f86f89e219" (UID: "c59d7f39-1709-49ba-8623-41f86f89e219"). InnerVolumeSpecName "kube-api-access-jqt6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.125628 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s" (OuterVolumeSpecName: "kube-api-access-rc65s") pod "565d5803-c324-4497-a928-a84f18aaea1b" (UID: "565d5803-c324-4497-a928-a84f18aaea1b"). InnerVolumeSpecName "kube-api-access-rc65s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.125983 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg" (OuterVolumeSpecName: "kube-api-access-264zg") pod "ab3d2425-2baf-4842-ac92-256766803f72" (UID: "ab3d2425-2baf-4842-ac92-256766803f72"). InnerVolumeSpecName "kube-api-access-264zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225731 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zckc\" (UniqueName: \"kubernetes.io/projected/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-kube-api-access-7zckc\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225763 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkbt8\" (UniqueName: \"kubernetes.io/projected/a7f689bc-82d9-4867-8505-8ccc6eb237ce-kube-api-access-gkbt8\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225772 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqt6g\" (UniqueName: \"kubernetes.io/projected/c59d7f39-1709-49ba-8623-41f86f89e219-kube-api-access-jqt6g\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225781 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc65s\" (UniqueName: \"kubernetes.io/projected/565d5803-c324-4497-a928-a84f18aaea1b-kube-api-access-rc65s\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225790 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7f689bc-82d9-4867-8505-8ccc6eb237ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225800 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d48bc89d-ef25-4b10-96f0-be3a93c6eff1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.225808 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-264zg\" (UniqueName: \"kubernetes.io/projected/ab3d2425-2baf-4842-ac92-256766803f72-kube-api-access-264zg\") on node \"crc\" DevicePath \"\"" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.439698 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-skvtb" event={"ID":"c59d7f39-1709-49ba-8623-41f86f89e219","Type":"ContainerDied","Data":"3873bb99beab6b3227e9593e3c15492efe8b0d9ed8e8593855048ba643c45b04"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.439755 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3873bb99beab6b3227e9593e3c15492efe8b0d9ed8e8593855048ba643c45b04" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.439708 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-skvtb" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.442334 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" event={"ID":"a7f689bc-82d9-4867-8505-8ccc6eb237ce","Type":"ContainerDied","Data":"42a7a9b900f3e2c5b1a453f133e79a25e6b73e0c481b8aff7e4bfd7771f3ffb2"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.442384 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42a7a9b900f3e2c5b1a453f133e79a25e6b73e0c481b8aff7e4bfd7771f3ffb2" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.442345 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0e7e-account-create-update-g6wnx" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.444510 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5657-account-create-update-tl2gr" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.444525 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5657-account-create-update-tl2gr" event={"ID":"ab3d2425-2baf-4842-ac92-256766803f72","Type":"ContainerDied","Data":"db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.444582 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db8e13508e30cc4a0ded4cbf4d0a3f5fb728a56ddfae537797d06afe3a821089" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.446478 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.446488 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8bcb-account-create-update-cf9b2" event={"ID":"d48bc89d-ef25-4b10-96f0-be3a93c6eff1","Type":"ContainerDied","Data":"9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.446552 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ddc857071a5973dd089b644130d6b04c40d1ba8e07e8f582020bc99b365a1ca" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.448915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-sgljh" event={"ID":"565d5803-c324-4497-a928-a84f18aaea1b","Type":"ContainerDied","Data":"bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.448956 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcb6d8d6c9c668f769f0d8ce3956ad0ff699e414ec26f61c792b0f1f77612097" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.449002 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-sgljh" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.451624 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-r6b5p" event={"ID":"fa6cd035-6852-4418-ac56-33fa1f5c2a45","Type":"ContainerDied","Data":"084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c"} Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.451652 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="084d4b5f770c460f7f7d0b6d9f176b50c36212b216bc72bb415a34a77b26e82c" Dec 12 08:22:50 crc kubenswrapper[4867]: I1212 08:22:50.451784 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-r6b5p" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.071934 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x2wpg"] Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073215 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48bc89d-ef25-4b10-96f0-be3a93c6eff1" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073286 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48bc89d-ef25-4b10-96f0-be3a93c6eff1" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073313 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3d2425-2baf-4842-ac92-256766803f72" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073327 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3d2425-2baf-4842-ac92-256766803f72" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073355 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c59d7f39-1709-49ba-8623-41f86f89e219" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073363 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c59d7f39-1709-49ba-8623-41f86f89e219" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073398 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f689bc-82d9-4867-8505-8ccc6eb237ce" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073407 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f689bc-82d9-4867-8505-8ccc6eb237ce" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073427 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565d5803-c324-4497-a928-a84f18aaea1b" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073435 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="565d5803-c324-4497-a928-a84f18aaea1b" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: E1212 08:22:52.073463 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6cd035-6852-4418-ac56-33fa1f5c2a45" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.073470 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6cd035-6852-4418-ac56-33fa1f5c2a45" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074017 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c59d7f39-1709-49ba-8623-41f86f89e219" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074052 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f689bc-82d9-4867-8505-8ccc6eb237ce" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074073 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48bc89d-ef25-4b10-96f0-be3a93c6eff1" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074101 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6cd035-6852-4418-ac56-33fa1f5c2a45" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074130 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3d2425-2baf-4842-ac92-256766803f72" containerName="mariadb-account-create-update" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.074159 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="565d5803-c324-4497-a928-a84f18aaea1b" containerName="mariadb-database-create" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.075207 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.079013 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.079663 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nqjws" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.079958 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.112684 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x2wpg"] Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.211833 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl799\" (UniqueName: \"kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.211929 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.212280 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.212440 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.314377 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.314596 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.314651 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.314732 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl799\" (UniqueName: \"kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.320985 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.321060 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.322261 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.335188 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl799\" (UniqueName: \"kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799\") pod \"nova-cell0-conductor-db-sync-x2wpg\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.410153 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:22:52 crc kubenswrapper[4867]: I1212 08:22:52.884690 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x2wpg"] Dec 12 08:22:53 crc kubenswrapper[4867]: I1212 08:22:53.646124 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" event={"ID":"ba8c7c36-4f4c-499c-9c6c-9eda87455841","Type":"ContainerStarted","Data":"db8d30491f10fea810da1b1255553f7dc9181544f73563c44abc9d6999259057"} Dec 12 08:22:53 crc kubenswrapper[4867]: I1212 08:22:53.949734 4867 scope.go:117] "RemoveContainer" containerID="bf5f85a7d63856957ff216c2e260e6944a782dc4e46a7c8f17c7e7152793090d" Dec 12 08:22:53 crc kubenswrapper[4867]: I1212 08:22:53.973806 4867 scope.go:117] "RemoveContainer" containerID="dedf6ca4ad16f45eb6bc92f2324981d45457f8ecc3f587f476394ca47e89cb58" Dec 12 08:22:53 crc kubenswrapper[4867]: I1212 08:22:53.999008 4867 scope.go:117] "RemoveContainer" containerID="b70bca5053c6d4ed48d8a30e799ddbd6632b5b3e4e0f73f8f3fd8c01accc596c" Dec 12 08:22:54 crc kubenswrapper[4867]: I1212 08:22:54.026355 4867 scope.go:117] "RemoveContainer" containerID="779c779d65e05adf36356069eb5181278aee7aa5aa8451a70ffd22354874090b" Dec 12 08:22:54 crc kubenswrapper[4867]: I1212 08:22:54.074420 4867 scope.go:117] "RemoveContainer" containerID="1688f4f05e2f518d285a162d40cc81d96ec60b5b606dbbc05c8ee927c834405c" Dec 12 08:23:02 crc kubenswrapper[4867]: I1212 08:23:02.730329 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" event={"ID":"ba8c7c36-4f4c-499c-9c6c-9eda87455841","Type":"ContainerStarted","Data":"f345dc8677ec0295f9398d424dfa621a62c935ec78b631f44f42a88720b24f12"} Dec 12 08:23:02 crc kubenswrapper[4867]: I1212 08:23:02.748412 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" podStartSLOduration=1.6779497 podStartE2EDuration="10.748394083s" podCreationTimestamp="2025-12-12 08:22:52 +0000 UTC" firstStartedPulling="2025-12-12 08:22:52.88609564 +0000 UTC m=+5660.457476909" lastFinishedPulling="2025-12-12 08:23:01.956540023 +0000 UTC m=+5669.527921292" observedRunningTime="2025-12-12 08:23:02.742687982 +0000 UTC m=+5670.314069271" watchObservedRunningTime="2025-12-12 08:23:02.748394083 +0000 UTC m=+5670.319775352" Dec 12 08:23:07 crc kubenswrapper[4867]: I1212 08:23:07.775759 4867 generic.go:334] "Generic (PLEG): container finished" podID="ba8c7c36-4f4c-499c-9c6c-9eda87455841" containerID="f345dc8677ec0295f9398d424dfa621a62c935ec78b631f44f42a88720b24f12" exitCode=0 Dec 12 08:23:07 crc kubenswrapper[4867]: I1212 08:23:07.775922 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" event={"ID":"ba8c7c36-4f4c-499c-9c6c-9eda87455841","Type":"ContainerDied","Data":"f345dc8677ec0295f9398d424dfa621a62c935ec78b631f44f42a88720b24f12"} Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.124572 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.261316 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle\") pod \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.261443 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl799\" (UniqueName: \"kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799\") pod \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.261646 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts\") pod \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.261703 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data\") pod \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\" (UID: \"ba8c7c36-4f4c-499c-9c6c-9eda87455841\") " Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.268512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799" (OuterVolumeSpecName: "kube-api-access-zl799") pod "ba8c7c36-4f4c-499c-9c6c-9eda87455841" (UID: "ba8c7c36-4f4c-499c-9c6c-9eda87455841"). InnerVolumeSpecName "kube-api-access-zl799". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.269100 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts" (OuterVolumeSpecName: "scripts") pod "ba8c7c36-4f4c-499c-9c6c-9eda87455841" (UID: "ba8c7c36-4f4c-499c-9c6c-9eda87455841"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.294650 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data" (OuterVolumeSpecName: "config-data") pod "ba8c7c36-4f4c-499c-9c6c-9eda87455841" (UID: "ba8c7c36-4f4c-499c-9c6c-9eda87455841"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.295094 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba8c7c36-4f4c-499c-9c6c-9eda87455841" (UID: "ba8c7c36-4f4c-499c-9c6c-9eda87455841"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.363903 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl799\" (UniqueName: \"kubernetes.io/projected/ba8c7c36-4f4c-499c-9c6c-9eda87455841-kube-api-access-zl799\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.365461 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.365511 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.365527 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8c7c36-4f4c-499c-9c6c-9eda87455841-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.795314 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" event={"ID":"ba8c7c36-4f4c-499c-9c6c-9eda87455841","Type":"ContainerDied","Data":"db8d30491f10fea810da1b1255553f7dc9181544f73563c44abc9d6999259057"} Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.795773 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db8d30491f10fea810da1b1255553f7dc9181544f73563c44abc9d6999259057" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.795370 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x2wpg" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.881565 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 08:23:09 crc kubenswrapper[4867]: E1212 08:23:09.881932 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba8c7c36-4f4c-499c-9c6c-9eda87455841" containerName="nova-cell0-conductor-db-sync" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.881948 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba8c7c36-4f4c-499c-9c6c-9eda87455841" containerName="nova-cell0-conductor-db-sync" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.882155 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba8c7c36-4f4c-499c-9c6c-9eda87455841" containerName="nova-cell0-conductor-db-sync" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.882799 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.885301 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.885908 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-nqjws" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.919150 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.975869 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.976152 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:09 crc kubenswrapper[4867]: I1212 08:23:09.976290 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnl55\" (UniqueName: \"kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.077844 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnl55\" (UniqueName: \"kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.077938 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.078037 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.083865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.084477 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.096145 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnl55\" (UniqueName: \"kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55\") pod \"nova-cell0-conductor-0\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.228375 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.690509 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 08:23:10 crc kubenswrapper[4867]: W1212 08:23:10.733493 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb556b6f3_e3b7_4459_9cd2_86c00b610173.slice/crio-36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359 WatchSource:0}: Error finding container 36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359: Status 404 returned error can't find the container with id 36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359 Dec 12 08:23:10 crc kubenswrapper[4867]: I1212 08:23:10.804506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b556b6f3-e3b7-4459-9cd2-86c00b610173","Type":"ContainerStarted","Data":"36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359"} Dec 12 08:23:11 crc kubenswrapper[4867]: I1212 08:23:11.815822 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b556b6f3-e3b7-4459-9cd2-86c00b610173","Type":"ContainerStarted","Data":"91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f"} Dec 12 08:23:11 crc kubenswrapper[4867]: I1212 08:23:11.816405 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:11 crc kubenswrapper[4867]: I1212 08:23:11.841610 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.841587201 podStartE2EDuration="2.841587201s" podCreationTimestamp="2025-12-12 08:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:11.831663355 +0000 UTC m=+5679.403044634" watchObservedRunningTime="2025-12-12 08:23:11.841587201 +0000 UTC m=+5679.412968470" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.260982 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.743978 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4g6xv"] Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.745102 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.747889 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.748266 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.756376 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4g6xv"] Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.886252 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.888036 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.892948 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.905196 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.905218 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.905510 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.905552 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8g85\" (UniqueName: \"kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.905704 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.922518 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.924110 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.928257 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 08:23:15 crc kubenswrapper[4867]: I1212 08:23:15.957105 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.007845 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.007929 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008034 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008106 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008212 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008266 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008386 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6cqd\" (UniqueName: \"kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008425 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crvj9\" (UniqueName: \"kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008531 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008564 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.008593 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8g85\" (UniqueName: \"kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.022129 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.025029 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.031842 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.051899 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.053639 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.058892 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8g85\" (UniqueName: \"kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85\") pod \"nova-cell0-cell-mapping-4g6xv\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.059577 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.070847 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.078484 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.111088 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112070 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112200 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112284 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112429 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112570 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6cqd\" (UniqueName: \"kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112613 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112660 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crvj9\" (UniqueName: \"kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.112901 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.117592 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.117855 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.118278 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.119857 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.120522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.125266 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.126281 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.147010 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crvj9\" (UniqueName: \"kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9\") pod \"nova-metadata-0\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.188931 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6cqd\" (UniqueName: \"kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd\") pod \"nova-api-0\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.196790 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.198089 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.205121 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.212416 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214091 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214318 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214435 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mpvw\" (UniqueName: \"kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214508 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214583 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214687 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214811 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.214875 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjpf2\" (UniqueName: \"kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.248929 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.259853 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.266904 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317022 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317278 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjpf2\" (UniqueName: \"kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317318 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317350 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317401 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317422 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mpvw\" (UniqueName: \"kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317437 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317460 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317492 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317511 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtxwk\" (UniqueName: \"kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.317529 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.318352 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.319988 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.321149 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.321396 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.330166 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.331742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.338608 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjpf2\" (UniqueName: \"kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.381440 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mpvw\" (UniqueName: \"kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw\") pod \"dnsmasq-dns-5fc5b649c9-fx44l\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.419298 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.419470 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtxwk\" (UniqueName: \"kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.419488 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.431176 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.432825 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.463045 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtxwk\" (UniqueName: \"kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk\") pod \"nova-scheduler-0\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.624091 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.644453 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.707492 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:16 crc kubenswrapper[4867]: W1212 08:23:16.878436 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c32d0b5_f42f_4361_b3c5_310a8b3108bb.slice/crio-db80fb07256eef13ea952511a4eb5b81f032fb2c75d54ca41dbbec3c07b3be80 WatchSource:0}: Error finding container db80fb07256eef13ea952511a4eb5b81f032fb2c75d54ca41dbbec3c07b3be80: Status 404 returned error can't find the container with id db80fb07256eef13ea952511a4eb5b81f032fb2c75d54ca41dbbec3c07b3be80 Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.881190 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:16 crc kubenswrapper[4867]: I1212 08:23:16.930458 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4g6xv"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.150210 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:17 crc kubenswrapper[4867]: W1212 08:23:17.166291 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48727be_70a1_4ee2_906e_b99847234198.slice/crio-733c056ddb81859c87f145b20124088ca4247e6582abd3579b5c10f15ca03b2b WatchSource:0}: Error finding container 733c056ddb81859c87f145b20124088ca4247e6582abd3579b5c10f15ca03b2b: Status 404 returned error can't find the container with id 733c056ddb81859c87f145b20124088ca4247e6582abd3579b5c10f15ca03b2b Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.308501 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.323814 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l749f"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.325688 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.328034 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.328731 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 12 08:23:17 crc kubenswrapper[4867]: W1212 08:23:17.335949 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod443ec657_a3c4_4787_a392_208a02028d88.slice/crio-54052b5653e046606e6895d1d62f4f9e8fab534ed29915a658951347a42b7b72 WatchSource:0}: Error finding container 54052b5653e046606e6895d1d62f4f9e8fab534ed29915a658951347a42b7b72: Status 404 returned error can't find the container with id 54052b5653e046606e6895d1d62f4f9e8fab534ed29915a658951347a42b7b72 Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.351564 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.361497 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l749f"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.427811 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.459742 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nnhd\" (UniqueName: \"kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.459916 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.459942 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.460110 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.561958 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.562015 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.562061 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.562114 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nnhd\" (UniqueName: \"kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.567289 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.567325 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.568068 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.590326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nnhd\" (UniqueName: \"kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd\") pod \"nova-cell1-conductor-db-sync-l749f\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.792209 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.950494 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"443ec657-a3c4-4787-a392-208a02028d88","Type":"ContainerStarted","Data":"54052b5653e046606e6895d1d62f4f9e8fab534ed29915a658951347a42b7b72"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.951974 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerStarted","Data":"733c056ddb81859c87f145b20124088ca4247e6582abd3579b5c10f15ca03b2b"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.955254 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4g6xv" event={"ID":"0a6d51e8-9cc1-48e6-95d4-afeea668e149","Type":"ContainerStarted","Data":"a6edc791ccee03374cf0bc0cf52daef7c1db7fa42c1ce88e510ec21d26bb2e3f"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.955284 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4g6xv" event={"ID":"0a6d51e8-9cc1-48e6-95d4-afeea668e149","Type":"ContainerStarted","Data":"3184c5247a0e1ef352da20fd57889f34903efdc15bc6774e7f69aa4426a12b8c"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.956713 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerStarted","Data":"db80fb07256eef13ea952511a4eb5b81f032fb2c75d54ca41dbbec3c07b3be80"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.958694 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eebaa5be-47cc-43fa-aba6-713c06a7df46","Type":"ContainerStarted","Data":"38d40b0bfe81ce3e734a02e262e1c6e0710717c4d1259ebaf100a8839546fda7"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.961698 4867 generic.go:334] "Generic (PLEG): container finished" podID="c33f7309-f9a3-4add-9887-749f61676452" containerID="0af4f965ee49f7ce8411b711a483551419065977c7b1ab1fa73d8f315ebe78ba" exitCode=0 Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.961752 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" event={"ID":"c33f7309-f9a3-4add-9887-749f61676452","Type":"ContainerDied","Data":"0af4f965ee49f7ce8411b711a483551419065977c7b1ab1fa73d8f315ebe78ba"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.961772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" event={"ID":"c33f7309-f9a3-4add-9887-749f61676452","Type":"ContainerStarted","Data":"16330b28bce0c748748de6f77ed84a14dec178e6a6f1314f499799df60243402"} Dec 12 08:23:17 crc kubenswrapper[4867]: I1212 08:23:17.990705 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4g6xv" podStartSLOduration=2.990681075 podStartE2EDuration="2.990681075s" podCreationTimestamp="2025-12-12 08:23:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:17.978565255 +0000 UTC m=+5685.549946534" watchObservedRunningTime="2025-12-12 08:23:17.990681075 +0000 UTC m=+5685.562062344" Dec 12 08:23:18 crc kubenswrapper[4867]: I1212 08:23:18.353130 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l749f"] Dec 12 08:23:18 crc kubenswrapper[4867]: W1212 08:23:18.382172 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17206bfe_043b_477a_9c1d_d66156fac47b.slice/crio-36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499 WatchSource:0}: Error finding container 36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499: Status 404 returned error can't find the container with id 36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499 Dec 12 08:23:18 crc kubenswrapper[4867]: I1212 08:23:18.972977 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l749f" event={"ID":"17206bfe-043b-477a-9c1d-d66156fac47b","Type":"ContainerStarted","Data":"36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499"} Dec 12 08:23:18 crc kubenswrapper[4867]: I1212 08:23:18.975691 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" event={"ID":"c33f7309-f9a3-4add-9887-749f61676452","Type":"ContainerStarted","Data":"527e851f4b1a85e49e4d67716cc580ab031c8f6e47294d3515ba32dccc5f94a8"} Dec 12 08:23:18 crc kubenswrapper[4867]: I1212 08:23:18.975884 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:18 crc kubenswrapper[4867]: I1212 08:23:18.998959 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" podStartSLOduration=2.998940042 podStartE2EDuration="2.998940042s" podCreationTimestamp="2025-12-12 08:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:18.994615425 +0000 UTC m=+5686.565996714" watchObservedRunningTime="2025-12-12 08:23:18.998940042 +0000 UTC m=+5686.570321311" Dec 12 08:23:19 crc kubenswrapper[4867]: I1212 08:23:19.948942 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:19 crc kubenswrapper[4867]: I1212 08:23:19.964033 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.067496 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-691b-account-create-update-8589s"] Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.077555 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-s95rc"] Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.091860 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-691b-account-create-update-8589s"] Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.101747 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-s95rc"] Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.852064 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce280bcc-4651-4777-8e50-f6951e781224" path="/var/lib/kubelet/pods/ce280bcc-4651-4777-8e50-f6951e781224/volumes" Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.856390 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbfa62db-247f-44d3-bd2b-efb127cb954e" path="/var/lib/kubelet/pods/fbfa62db-247f-44d3-bd2b-efb127cb954e/volumes" Dec 12 08:23:20 crc kubenswrapper[4867]: I1212 08:23:20.995408 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l749f" event={"ID":"17206bfe-043b-477a-9c1d-d66156fac47b","Type":"ContainerStarted","Data":"ad20c942d6a9d58e9927b8f68be8789211e06512e5690aaa3e25a1e31e624190"} Dec 12 08:23:21 crc kubenswrapper[4867]: I1212 08:23:21.014696 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-l749f" podStartSLOduration=4.014673549 podStartE2EDuration="4.014673549s" podCreationTimestamp="2025-12-12 08:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:21.012362371 +0000 UTC m=+5688.583743640" watchObservedRunningTime="2025-12-12 08:23:21.014673549 +0000 UTC m=+5688.586054818" Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.005275 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"443ec657-a3c4-4787-a392-208a02028d88","Type":"ContainerStarted","Data":"8bd764366adb88f964f492d480139f012ebb8ac8810e901bfb7238dd7f77da03"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.005983 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="443ec657-a3c4-4787-a392-208a02028d88" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8bd764366adb88f964f492d480139f012ebb8ac8810e901bfb7238dd7f77da03" gracePeriod=30 Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.006805 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerStarted","Data":"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.006917 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerStarted","Data":"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.009542 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerStarted","Data":"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.009583 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerStarted","Data":"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.012418 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eebaa5be-47cc-43fa-aba6-713c06a7df46","Type":"ContainerStarted","Data":"7f841a1b3956dd375cc5b9c116e73d4359660288330f8b616f58f4257e61a561"} Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.025477 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.638801591 podStartE2EDuration="7.025462109s" podCreationTimestamp="2025-12-12 08:23:15 +0000 UTC" firstStartedPulling="2025-12-12 08:23:17.34415874 +0000 UTC m=+5684.915540009" lastFinishedPulling="2025-12-12 08:23:20.730819258 +0000 UTC m=+5688.302200527" observedRunningTime="2025-12-12 08:23:22.023588393 +0000 UTC m=+5689.594969662" watchObservedRunningTime="2025-12-12 08:23:22.025462109 +0000 UTC m=+5689.596843378" Dec 12 08:23:22 crc kubenswrapper[4867]: I1212 08:23:22.042856 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.747572859 podStartE2EDuration="6.04283995s" podCreationTimestamp="2025-12-12 08:23:16 +0000 UTC" firstStartedPulling="2025-12-12 08:23:17.435620288 +0000 UTC m=+5685.007001557" lastFinishedPulling="2025-12-12 08:23:20.730887379 +0000 UTC m=+5688.302268648" observedRunningTime="2025-12-12 08:23:22.040586074 +0000 UTC m=+5689.611967343" watchObservedRunningTime="2025-12-12 08:23:22.04283995 +0000 UTC m=+5689.614221209" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.018523 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-log" containerID="cri-o://4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" gracePeriod=30 Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.019098 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-metadata" containerID="cri-o://5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" gracePeriod=30 Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.044896 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.485549892 podStartE2EDuration="8.044874843s" podCreationTimestamp="2025-12-12 08:23:15 +0000 UTC" firstStartedPulling="2025-12-12 08:23:17.173699132 +0000 UTC m=+5684.745080401" lastFinishedPulling="2025-12-12 08:23:20.733024083 +0000 UTC m=+5688.304405352" observedRunningTime="2025-12-12 08:23:23.038152516 +0000 UTC m=+5690.609533805" watchObservedRunningTime="2025-12-12 08:23:23.044874843 +0000 UTC m=+5690.616256102" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.090765 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.316234443 podStartE2EDuration="8.090743191s" podCreationTimestamp="2025-12-12 08:23:15 +0000 UTC" firstStartedPulling="2025-12-12 08:23:16.958974266 +0000 UTC m=+5684.530355535" lastFinishedPulling="2025-12-12 08:23:20.733482974 +0000 UTC m=+5688.304864283" observedRunningTime="2025-12-12 08:23:23.082839045 +0000 UTC m=+5690.654220384" watchObservedRunningTime="2025-12-12 08:23:23.090743191 +0000 UTC m=+5690.662124460" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.630996 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.695112 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle\") pod \"c48727be-70a1-4ee2-906e-b99847234198\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.695191 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data\") pod \"c48727be-70a1-4ee2-906e-b99847234198\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.695217 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crvj9\" (UniqueName: \"kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9\") pod \"c48727be-70a1-4ee2-906e-b99847234198\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.695318 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs\") pod \"c48727be-70a1-4ee2-906e-b99847234198\" (UID: \"c48727be-70a1-4ee2-906e-b99847234198\") " Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.695928 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs" (OuterVolumeSpecName: "logs") pod "c48727be-70a1-4ee2-906e-b99847234198" (UID: "c48727be-70a1-4ee2-906e-b99847234198"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.700827 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9" (OuterVolumeSpecName: "kube-api-access-crvj9") pod "c48727be-70a1-4ee2-906e-b99847234198" (UID: "c48727be-70a1-4ee2-906e-b99847234198"). InnerVolumeSpecName "kube-api-access-crvj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.727809 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c48727be-70a1-4ee2-906e-b99847234198" (UID: "c48727be-70a1-4ee2-906e-b99847234198"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.729607 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data" (OuterVolumeSpecName: "config-data") pod "c48727be-70a1-4ee2-906e-b99847234198" (UID: "c48727be-70a1-4ee2-906e-b99847234198"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.799771 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.799819 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c48727be-70a1-4ee2-906e-b99847234198-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.799829 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crvj9\" (UniqueName: \"kubernetes.io/projected/c48727be-70a1-4ee2-906e-b99847234198-kube-api-access-crvj9\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:23 crc kubenswrapper[4867]: I1212 08:23:23.799843 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c48727be-70a1-4ee2-906e-b99847234198-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.028697 4867 generic.go:334] "Generic (PLEG): container finished" podID="c48727be-70a1-4ee2-906e-b99847234198" containerID="5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" exitCode=0 Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.028992 4867 generic.go:334] "Generic (PLEG): container finished" podID="c48727be-70a1-4ee2-906e-b99847234198" containerID="4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" exitCode=143 Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.028761 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.028759 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerDied","Data":"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed"} Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.030144 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerDied","Data":"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb"} Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.030190 4867 scope.go:117] "RemoveContainer" containerID="5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.030192 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c48727be-70a1-4ee2-906e-b99847234198","Type":"ContainerDied","Data":"733c056ddb81859c87f145b20124088ca4247e6582abd3579b5c10f15ca03b2b"} Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.032011 4867 generic.go:334] "Generic (PLEG): container finished" podID="0a6d51e8-9cc1-48e6-95d4-afeea668e149" containerID="a6edc791ccee03374cf0bc0cf52daef7c1db7fa42c1ce88e510ec21d26bb2e3f" exitCode=0 Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.032113 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4g6xv" event={"ID":"0a6d51e8-9cc1-48e6-95d4-afeea668e149","Type":"ContainerDied","Data":"a6edc791ccee03374cf0bc0cf52daef7c1db7fa42c1ce88e510ec21d26bb2e3f"} Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.034552 4867 generic.go:334] "Generic (PLEG): container finished" podID="17206bfe-043b-477a-9c1d-d66156fac47b" containerID="ad20c942d6a9d58e9927b8f68be8789211e06512e5690aaa3e25a1e31e624190" exitCode=0 Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.034615 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l749f" event={"ID":"17206bfe-043b-477a-9c1d-d66156fac47b","Type":"ContainerDied","Data":"ad20c942d6a9d58e9927b8f68be8789211e06512e5690aaa3e25a1e31e624190"} Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.053529 4867 scope.go:117] "RemoveContainer" containerID="4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.075410 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.081739 4867 scope.go:117] "RemoveContainer" containerID="5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" Dec 12 08:23:24 crc kubenswrapper[4867]: E1212 08:23:24.082619 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed\": container with ID starting with 5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed not found: ID does not exist" containerID="5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.082683 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed"} err="failed to get container status \"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed\": rpc error: code = NotFound desc = could not find container \"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed\": container with ID starting with 5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed not found: ID does not exist" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.082718 4867 scope.go:117] "RemoveContainer" containerID="4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" Dec 12 08:23:24 crc kubenswrapper[4867]: E1212 08:23:24.083045 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb\": container with ID starting with 4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb not found: ID does not exist" containerID="4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.083073 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb"} err="failed to get container status \"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb\": rpc error: code = NotFound desc = could not find container \"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb\": container with ID starting with 4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb not found: ID does not exist" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.083091 4867 scope.go:117] "RemoveContainer" containerID="5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.086293 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.087465 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed"} err="failed to get container status \"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed\": rpc error: code = NotFound desc = could not find container \"5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed\": container with ID starting with 5afb0a8097fd9372e26663642532a3f5359428ba38cffb0e81b6636d8b9119ed not found: ID does not exist" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.087523 4867 scope.go:117] "RemoveContainer" containerID="4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.087988 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb"} err="failed to get container status \"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb\": rpc error: code = NotFound desc = could not find container \"4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb\": container with ID starting with 4908c21a344fee637418c31c64ea3b3c2123009c5869fbe63c4b14f88a913dbb not found: ID does not exist" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.115739 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:24 crc kubenswrapper[4867]: E1212 08:23:24.116296 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-log" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.116317 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-log" Dec 12 08:23:24 crc kubenswrapper[4867]: E1212 08:23:24.116329 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-metadata" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.116336 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-metadata" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.116545 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-log" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.116567 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48727be-70a1-4ee2-906e-b99847234198" containerName="nova-metadata-metadata" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.129818 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.129920 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.132851 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.133048 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.207986 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.208064 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.208363 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.208424 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.208588 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l9xr\" (UniqueName: \"kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.310461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l9xr\" (UniqueName: \"kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.310558 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.310598 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.310679 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.310696 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.311168 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.314997 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.315742 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.316287 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.332586 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l9xr\" (UniqueName: \"kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr\") pod \"nova-metadata-0\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.496470 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.849512 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48727be-70a1-4ee2-906e-b99847234198" path="/var/lib/kubelet/pods/c48727be-70a1-4ee2-906e-b99847234198/volumes" Dec 12 08:23:24 crc kubenswrapper[4867]: I1212 08:23:24.967428 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:24 crc kubenswrapper[4867]: W1212 08:23:24.973038 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78dc37b1_00b7_434b_95a9_337c1a6e83b6.slice/crio-60c5397b8b8bdd9419fe0ce92fed07beacc67745d40d5e8dc9f1302e8bf7bd25 WatchSource:0}: Error finding container 60c5397b8b8bdd9419fe0ce92fed07beacc67745d40d5e8dc9f1302e8bf7bd25: Status 404 returned error can't find the container with id 60c5397b8b8bdd9419fe0ce92fed07beacc67745d40d5e8dc9f1302e8bf7bd25 Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.044117 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerStarted","Data":"60c5397b8b8bdd9419fe0ce92fed07beacc67745d40d5e8dc9f1302e8bf7bd25"} Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.421300 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.426741 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547096 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data\") pod \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547503 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts\") pod \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547636 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nnhd\" (UniqueName: \"kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd\") pod \"17206bfe-043b-477a-9c1d-d66156fac47b\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547673 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle\") pod \"17206bfe-043b-477a-9c1d-d66156fac47b\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547733 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data\") pod \"17206bfe-043b-477a-9c1d-d66156fac47b\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.547775 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle\") pod \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.548031 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts\") pod \"17206bfe-043b-477a-9c1d-d66156fac47b\" (UID: \"17206bfe-043b-477a-9c1d-d66156fac47b\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.548407 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8g85\" (UniqueName: \"kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85\") pod \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\" (UID: \"0a6d51e8-9cc1-48e6-95d4-afeea668e149\") " Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.551213 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd" (OuterVolumeSpecName: "kube-api-access-9nnhd") pod "17206bfe-043b-477a-9c1d-d66156fac47b" (UID: "17206bfe-043b-477a-9c1d-d66156fac47b"). InnerVolumeSpecName "kube-api-access-9nnhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.551682 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts" (OuterVolumeSpecName: "scripts") pod "17206bfe-043b-477a-9c1d-d66156fac47b" (UID: "17206bfe-043b-477a-9c1d-d66156fac47b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.551833 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85" (OuterVolumeSpecName: "kube-api-access-c8g85") pod "0a6d51e8-9cc1-48e6-95d4-afeea668e149" (UID: "0a6d51e8-9cc1-48e6-95d4-afeea668e149"). InnerVolumeSpecName "kube-api-access-c8g85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.551839 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts" (OuterVolumeSpecName: "scripts") pod "0a6d51e8-9cc1-48e6-95d4-afeea668e149" (UID: "0a6d51e8-9cc1-48e6-95d4-afeea668e149"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.573392 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data" (OuterVolumeSpecName: "config-data") pod "0a6d51e8-9cc1-48e6-95d4-afeea668e149" (UID: "0a6d51e8-9cc1-48e6-95d4-afeea668e149"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.574972 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data" (OuterVolumeSpecName: "config-data") pod "17206bfe-043b-477a-9c1d-d66156fac47b" (UID: "17206bfe-043b-477a-9c1d-d66156fac47b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.575900 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a6d51e8-9cc1-48e6-95d4-afeea668e149" (UID: "0a6d51e8-9cc1-48e6-95d4-afeea668e149"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.586930 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17206bfe-043b-477a-9c1d-d66156fac47b" (UID: "17206bfe-043b-477a-9c1d-d66156fac47b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650036 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650067 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8g85\" (UniqueName: \"kubernetes.io/projected/0a6d51e8-9cc1-48e6-95d4-afeea668e149-kube-api-access-c8g85\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650082 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650090 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650098 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nnhd\" (UniqueName: \"kubernetes.io/projected/17206bfe-043b-477a-9c1d-d66156fac47b-kube-api-access-9nnhd\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650127 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650135 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17206bfe-043b-477a-9c1d-d66156fac47b-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:25 crc kubenswrapper[4867]: I1212 08:23:25.650143 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6d51e8-9cc1-48e6-95d4-afeea668e149-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.054753 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l749f" event={"ID":"17206bfe-043b-477a-9c1d-d66156fac47b","Type":"ContainerDied","Data":"36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499"} Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.054774 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l749f" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.054796 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36d9a943992340dcb4d6e42437d96f11a01bd7a20078e13477b9ccb816f6e499" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.056877 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerStarted","Data":"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3"} Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.056927 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerStarted","Data":"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf"} Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.058393 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4g6xv" event={"ID":"0a6d51e8-9cc1-48e6-95d4-afeea668e149","Type":"ContainerDied","Data":"3184c5247a0e1ef352da20fd57889f34903efdc15bc6774e7f69aa4426a12b8c"} Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.058514 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3184c5247a0e1ef352da20fd57889f34903efdc15bc6774e7f69aa4426a12b8c" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.058620 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4g6xv" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.084777 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.084755121 podStartE2EDuration="2.084755121s" podCreationTimestamp="2025-12-12 08:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:26.074021265 +0000 UTC m=+5693.645402544" watchObservedRunningTime="2025-12-12 08:23:26.084755121 +0000 UTC m=+5693.656136390" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.181696 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 08:23:26 crc kubenswrapper[4867]: E1212 08:23:26.182188 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17206bfe-043b-477a-9c1d-d66156fac47b" containerName="nova-cell1-conductor-db-sync" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.182205 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="17206bfe-043b-477a-9c1d-d66156fac47b" containerName="nova-cell1-conductor-db-sync" Dec 12 08:23:26 crc kubenswrapper[4867]: E1212 08:23:26.182268 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a6d51e8-9cc1-48e6-95d4-afeea668e149" containerName="nova-manage" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.182277 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a6d51e8-9cc1-48e6-95d4-afeea668e149" containerName="nova-manage" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.182930 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a6d51e8-9cc1-48e6-95d4-afeea668e149" containerName="nova-manage" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.182960 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="17206bfe-043b-477a-9c1d-d66156fac47b" containerName="nova-cell1-conductor-db-sync" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.186809 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.189186 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.206208 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.206435 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.209826 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.264699 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2896\" (UniqueName: \"kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.264756 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.264820 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.276371 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.289342 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.289587 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="eebaa5be-47cc-43fa-aba6-713c06a7df46" containerName="nova-scheduler-scheduler" containerID="cri-o://7f841a1b3956dd375cc5b9c116e73d4359660288330f8b616f58f4257e61a561" gracePeriod=30 Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.293203 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.366287 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2896\" (UniqueName: \"kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.366336 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.366386 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.370370 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.370553 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.383980 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2896\" (UniqueName: \"kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896\") pod \"nova-cell1-conductor-0\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.552679 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.624819 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.646516 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.708947 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.757988 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:23:26 crc kubenswrapper[4867]: I1212 08:23:26.758290 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="dnsmasq-dns" containerID="cri-o://dfc462fba9233854bf0c33db5ad6610b989987dcdb0f2333591adfebbc9b788d" gracePeriod=10 Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.080442 4867 generic.go:334] "Generic (PLEG): container finished" podID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerID="dfc462fba9233854bf0c33db5ad6610b989987dcdb0f2333591adfebbc9b788d" exitCode=0 Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.081146 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" event={"ID":"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979","Type":"ContainerDied","Data":"dfc462fba9233854bf0c33db5ad6610b989987dcdb0f2333591adfebbc9b788d"} Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.081331 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-log" containerID="cri-o://3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc" gracePeriod=30 Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.081468 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-api" containerID="cri-o://fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240" gracePeriod=30 Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.095372 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": EOF" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.095548 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": EOF" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.253426 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.428528 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.491175 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc\") pod \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.491292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb\") pod \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.491319 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb\") pod \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.491401 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config\") pod \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.491450 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwbln\" (UniqueName: \"kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln\") pod \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\" (UID: \"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979\") " Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.504579 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln" (OuterVolumeSpecName: "kube-api-access-bwbln") pod "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" (UID: "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979"). InnerVolumeSpecName "kube-api-access-bwbln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.565997 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" (UID: "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.575349 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config" (OuterVolumeSpecName: "config") pod "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" (UID: "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.583848 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" (UID: "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.593303 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" (UID: "5bc9fb1f-750c-46d8-8a5a-548d4d5d4979"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.595388 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.595493 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.595578 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.595651 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:27 crc kubenswrapper[4867]: I1212 08:23:27.595740 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwbln\" (UniqueName: \"kubernetes.io/projected/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979-kube-api-access-bwbln\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.091090 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3481dcf-092f-46a3-a41a-81f655801ddb","Type":"ContainerStarted","Data":"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00"} Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.091168 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3481dcf-092f-46a3-a41a-81f655801ddb","Type":"ContainerStarted","Data":"2b6bad0fb713b173042c0773d6bd1279aa51c1f5d30e64e2834fabf3636f0220"} Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.091188 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.093121 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerID="3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc" exitCode=143 Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.093197 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerDied","Data":"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc"} Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.096356 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-log" containerID="cri-o://29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" gracePeriod=30 Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.096612 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.097115 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95c76c545-ttnbc" event={"ID":"5bc9fb1f-750c-46d8-8a5a-548d4d5d4979","Type":"ContainerDied","Data":"cf01d8d82a00d58f7006e34f961af28ad3931872ede85edd7e38059b05b6cd0f"} Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.097249 4867 scope.go:117] "RemoveContainer" containerID="dfc462fba9233854bf0c33db5ad6610b989987dcdb0f2333591adfebbc9b788d" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.097537 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-metadata" containerID="cri-o://5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" gracePeriod=30 Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.125018 4867 scope.go:117] "RemoveContainer" containerID="663ce92697c58fd77df4f91db759ad92c6f4bd238864646cba8f2221813208a5" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.128497 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.128473621 podStartE2EDuration="2.128473621s" podCreationTimestamp="2025-12-12 08:23:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:28.114905744 +0000 UTC m=+5695.686287033" watchObservedRunningTime="2025-12-12 08:23:28.128473621 +0000 UTC m=+5695.699854900" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.233695 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.242257 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-95c76c545-ttnbc"] Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.647511 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.730577 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l9xr\" (UniqueName: \"kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr\") pod \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.731789 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle\") pod \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.732137 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs\") pod \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.732272 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data\") pod \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.732320 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs\") pod \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\" (UID: \"78dc37b1-00b7-434b-95a9-337c1a6e83b6\") " Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.732821 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs" (OuterVolumeSpecName: "logs") pod "78dc37b1-00b7-434b-95a9-337c1a6e83b6" (UID: "78dc37b1-00b7-434b-95a9-337c1a6e83b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.733194 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78dc37b1-00b7-434b-95a9-337c1a6e83b6-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.735247 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr" (OuterVolumeSpecName: "kube-api-access-2l9xr") pod "78dc37b1-00b7-434b-95a9-337c1a6e83b6" (UID: "78dc37b1-00b7-434b-95a9-337c1a6e83b6"). InnerVolumeSpecName "kube-api-access-2l9xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.757308 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data" (OuterVolumeSpecName: "config-data") pod "78dc37b1-00b7-434b-95a9-337c1a6e83b6" (UID: "78dc37b1-00b7-434b-95a9-337c1a6e83b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.763051 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78dc37b1-00b7-434b-95a9-337c1a6e83b6" (UID: "78dc37b1-00b7-434b-95a9-337c1a6e83b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.789821 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "78dc37b1-00b7-434b-95a9-337c1a6e83b6" (UID: "78dc37b1-00b7-434b-95a9-337c1a6e83b6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.834499 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l9xr\" (UniqueName: \"kubernetes.io/projected/78dc37b1-00b7-434b-95a9-337c1a6e83b6-kube-api-access-2l9xr\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.834534 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.834543 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.834554 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78dc37b1-00b7-434b-95a9-337c1a6e83b6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:28 crc kubenswrapper[4867]: I1212 08:23:28.849975 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" path="/var/lib/kubelet/pods/5bc9fb1f-750c-46d8-8a5a-548d4d5d4979/volumes" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106324 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106220 4867 generic.go:334] "Generic (PLEG): container finished" podID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerID="5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" exitCode=0 Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106773 4867 generic.go:334] "Generic (PLEG): container finished" podID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerID="29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" exitCode=143 Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106362 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerDied","Data":"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3"} Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106912 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerDied","Data":"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf"} Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106938 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78dc37b1-00b7-434b-95a9-337c1a6e83b6","Type":"ContainerDied","Data":"60c5397b8b8bdd9419fe0ce92fed07beacc67745d40d5e8dc9f1302e8bf7bd25"} Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.106957 4867 scope.go:117] "RemoveContainer" containerID="5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.133512 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.157757 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161462 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161785 4867 scope.go:117] "RemoveContainer" containerID="29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.161886 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="dnsmasq-dns" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161908 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="dnsmasq-dns" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.161924 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-log" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161931 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-log" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.161943 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-metadata" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161951 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-metadata" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.161983 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="init" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.161991 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="init" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.162336 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-log" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.162349 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" containerName="nova-metadata-metadata" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.162369 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc9fb1f-750c-46d8-8a5a-548d4d5d4979" containerName="dnsmasq-dns" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.163496 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.170329 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.170561 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.208747 4867 scope.go:117] "RemoveContainer" containerID="5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.211598 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3\": container with ID starting with 5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3 not found: ID does not exist" containerID="5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.211641 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3"} err="failed to get container status \"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3\": rpc error: code = NotFound desc = could not find container \"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3\": container with ID starting with 5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3 not found: ID does not exist" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.211667 4867 scope.go:117] "RemoveContainer" containerID="29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" Dec 12 08:23:29 crc kubenswrapper[4867]: E1212 08:23:29.212121 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf\": container with ID starting with 29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf not found: ID does not exist" containerID="29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.212148 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf"} err="failed to get container status \"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf\": rpc error: code = NotFound desc = could not find container \"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf\": container with ID starting with 29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf not found: ID does not exist" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.212168 4867 scope.go:117] "RemoveContainer" containerID="5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.215977 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3"} err="failed to get container status \"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3\": rpc error: code = NotFound desc = could not find container \"5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3\": container with ID starting with 5a354810ae8c03ba7702db41d928099e23b7aaab3b29dc5b7052335a4e7918f3 not found: ID does not exist" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.215999 4867 scope.go:117] "RemoveContainer" containerID="29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.216055 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.219426 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf"} err="failed to get container status \"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf\": rpc error: code = NotFound desc = could not find container \"29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf\": container with ID starting with 29861b0d33c906ca860f71f1630c71f86e6e088ac48df65f1d33bfd26fdc65cf not found: ID does not exist" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.248739 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk5rq\" (UniqueName: \"kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.248836 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.248866 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.248884 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.248950 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.351429 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk5rq\" (UniqueName: \"kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.351520 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.351585 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.351613 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.351729 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.352279 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.356624 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.356673 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.357677 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.367099 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk5rq\" (UniqueName: \"kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq\") pod \"nova-metadata-0\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.513086 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:23:29 crc kubenswrapper[4867]: I1212 08:23:29.979427 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:23:30 crc kubenswrapper[4867]: I1212 08:23:30.121114 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerStarted","Data":"401b82efcba1189838bf526be7a3382491d14f7625ff3db5b76cd4e9b3987366"} Dec 12 08:23:30 crc kubenswrapper[4867]: I1212 08:23:30.850921 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78dc37b1-00b7-434b-95a9-337c1a6e83b6" path="/var/lib/kubelet/pods/78dc37b1-00b7-434b-95a9-337c1a6e83b6/volumes" Dec 12 08:23:31 crc kubenswrapper[4867]: I1212 08:23:31.133977 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerStarted","Data":"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9"} Dec 12 08:23:31 crc kubenswrapper[4867]: I1212 08:23:31.134026 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerStarted","Data":"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2"} Dec 12 08:23:31 crc kubenswrapper[4867]: I1212 08:23:31.156515 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.156493455 podStartE2EDuration="2.156493455s" podCreationTimestamp="2025-12-12 08:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:31.151261885 +0000 UTC m=+5698.722643164" watchObservedRunningTime="2025-12-12 08:23:31.156493455 +0000 UTC m=+5698.727874724" Dec 12 08:23:32 crc kubenswrapper[4867]: I1212 08:23:32.052258 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-tgxgc"] Dec 12 08:23:32 crc kubenswrapper[4867]: I1212 08:23:32.062815 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-tgxgc"] Dec 12 08:23:32 crc kubenswrapper[4867]: I1212 08:23:32.847809 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8915409-bbc7-4db4-9a74-0bbfe956f80f" path="/var/lib/kubelet/pods/f8915409-bbc7-4db4-9a74-0bbfe956f80f/volumes" Dec 12 08:23:34 crc kubenswrapper[4867]: I1212 08:23:34.514033 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 08:23:34 crc kubenswrapper[4867]: I1212 08:23:34.514412 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 08:23:36 crc kubenswrapper[4867]: I1212 08:23:36.584857 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 12 08:23:39 crc kubenswrapper[4867]: I1212 08:23:39.513526 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 08:23:39 crc kubenswrapper[4867]: I1212 08:23:39.514091 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 08:23:40 crc kubenswrapper[4867]: I1212 08:23:40.530501 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:40 crc kubenswrapper[4867]: I1212 08:23:40.530501 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.023302 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.157869 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle\") pod \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.158695 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data\") pod \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.158913 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs\") pod \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.158945 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6cqd\" (UniqueName: \"kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd\") pod \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\" (UID: \"8c32d0b5-f42f-4361-b3c5-310a8b3108bb\") " Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.159540 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs" (OuterVolumeSpecName: "logs") pod "8c32d0b5-f42f-4361-b3c5-310a8b3108bb" (UID: "8c32d0b5-f42f-4361-b3c5-310a8b3108bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.160027 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.165328 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd" (OuterVolumeSpecName: "kube-api-access-n6cqd") pod "8c32d0b5-f42f-4361-b3c5-310a8b3108bb" (UID: "8c32d0b5-f42f-4361-b3c5-310a8b3108bb"). InnerVolumeSpecName "kube-api-access-n6cqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.192216 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c32d0b5-f42f-4361-b3c5-310a8b3108bb" (UID: "8c32d0b5-f42f-4361-b3c5-310a8b3108bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.220699 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data" (OuterVolumeSpecName: "config-data") pod "8c32d0b5-f42f-4361-b3c5-310a8b3108bb" (UID: "8c32d0b5-f42f-4361-b3c5-310a8b3108bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.224902 4867 generic.go:334] "Generic (PLEG): container finished" podID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerID="fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240" exitCode=0 Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.224958 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerDied","Data":"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240"} Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.224991 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8c32d0b5-f42f-4361-b3c5-310a8b3108bb","Type":"ContainerDied","Data":"db80fb07256eef13ea952511a4eb5b81f032fb2c75d54ca41dbbec3c07b3be80"} Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.225011 4867 scope.go:117] "RemoveContainer" containerID="fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.225189 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.262182 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6cqd\" (UniqueName: \"kubernetes.io/projected/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-kube-api-access-n6cqd\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.262363 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.262428 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c32d0b5-f42f-4361-b3c5-310a8b3108bb-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.282168 4867 scope.go:117] "RemoveContainer" containerID="3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.294381 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.314304 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.329805 4867 scope.go:117] "RemoveContainer" containerID="fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240" Dec 12 08:23:41 crc kubenswrapper[4867]: E1212 08:23:41.331032 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240\": container with ID starting with fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240 not found: ID does not exist" containerID="fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.331064 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240"} err="failed to get container status \"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240\": rpc error: code = NotFound desc = could not find container \"fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240\": container with ID starting with fa811a90b64982184c16ec67382072222a6e29cdf88a1db51615734775529240 not found: ID does not exist" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.331090 4867 scope.go:117] "RemoveContainer" containerID="3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc" Dec 12 08:23:41 crc kubenswrapper[4867]: E1212 08:23:41.331550 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc\": container with ID starting with 3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc not found: ID does not exist" containerID="3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.331629 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc"} err="failed to get container status \"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc\": rpc error: code = NotFound desc = could not find container \"3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc\": container with ID starting with 3e61d8321cd86d5de64e3503daa0aef7abf3a59e6cebd0db4450d01ec546cabc not found: ID does not exist" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.333278 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:41 crc kubenswrapper[4867]: E1212 08:23:41.333700 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-log" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.333716 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-log" Dec 12 08:23:41 crc kubenswrapper[4867]: E1212 08:23:41.333740 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-api" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.333746 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-api" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.333926 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-log" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.333934 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" containerName="nova-api-api" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.335600 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.338514 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.348563 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.470289 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.470358 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.470383 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.470655 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hdtv\" (UniqueName: \"kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.572097 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hdtv\" (UniqueName: \"kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.572415 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.572470 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.572497 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.572933 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.577293 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.588652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.595653 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hdtv\" (UniqueName: \"kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv\") pod \"nova-api-0\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " pod="openstack/nova-api-0" Dec 12 08:23:41 crc kubenswrapper[4867]: I1212 08:23:41.654476 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:23:42 crc kubenswrapper[4867]: I1212 08:23:42.165801 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:23:42 crc kubenswrapper[4867]: I1212 08:23:42.237770 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerStarted","Data":"f25d2203a5d6d4c20face71539694536414624b04119b34fac5ac193992afff3"} Dec 12 08:23:42 crc kubenswrapper[4867]: I1212 08:23:42.869051 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c32d0b5-f42f-4361-b3c5-310a8b3108bb" path="/var/lib/kubelet/pods/8c32d0b5-f42f-4361-b3c5-310a8b3108bb/volumes" Dec 12 08:23:43 crc kubenswrapper[4867]: I1212 08:23:43.249884 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerStarted","Data":"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1"} Dec 12 08:23:43 crc kubenswrapper[4867]: I1212 08:23:43.250489 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerStarted","Data":"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9"} Dec 12 08:23:43 crc kubenswrapper[4867]: I1212 08:23:43.274264 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.274217789 podStartE2EDuration="2.274217789s" podCreationTimestamp="2025-12-12 08:23:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:43.272874085 +0000 UTC m=+5710.844255394" watchObservedRunningTime="2025-12-12 08:23:43.274217789 +0000 UTC m=+5710.845599068" Dec 12 08:23:45 crc kubenswrapper[4867]: I1212 08:23:45.026591 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bfpsl"] Dec 12 08:23:45 crc kubenswrapper[4867]: I1212 08:23:45.035272 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bfpsl"] Dec 12 08:23:46 crc kubenswrapper[4867]: I1212 08:23:46.849857 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fc25c26-edca-4bc5-861a-f66e643cf7f0" path="/var/lib/kubelet/pods/0fc25c26-edca-4bc5-861a-f66e643cf7f0/volumes" Dec 12 08:23:50 crc kubenswrapper[4867]: I1212 08:23:50.521460 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:50 crc kubenswrapper[4867]: I1212 08:23:50.521495 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:51 crc kubenswrapper[4867]: I1212 08:23:51.655089 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:23:51 crc kubenswrapper[4867]: I1212 08:23:51.655144 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.355688 4867 generic.go:334] "Generic (PLEG): container finished" podID="443ec657-a3c4-4787-a392-208a02028d88" containerID="8bd764366adb88f964f492d480139f012ebb8ac8810e901bfb7238dd7f77da03" exitCode=137 Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.356558 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"443ec657-a3c4-4787-a392-208a02028d88","Type":"ContainerDied","Data":"8bd764366adb88f964f492d480139f012ebb8ac8810e901bfb7238dd7f77da03"} Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.490404 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.626424 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data\") pod \"443ec657-a3c4-4787-a392-208a02028d88\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.626514 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle\") pod \"443ec657-a3c4-4787-a392-208a02028d88\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.626640 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjpf2\" (UniqueName: \"kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2\") pod \"443ec657-a3c4-4787-a392-208a02028d88\" (UID: \"443ec657-a3c4-4787-a392-208a02028d88\") " Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.632536 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2" (OuterVolumeSpecName: "kube-api-access-wjpf2") pod "443ec657-a3c4-4787-a392-208a02028d88" (UID: "443ec657-a3c4-4787-a392-208a02028d88"). InnerVolumeSpecName "kube-api-access-wjpf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.669317 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data" (OuterVolumeSpecName: "config-data") pod "443ec657-a3c4-4787-a392-208a02028d88" (UID: "443ec657-a3c4-4787-a392-208a02028d88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.675674 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "443ec657-a3c4-4787-a392-208a02028d88" (UID: "443ec657-a3c4-4787-a392-208a02028d88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.728686 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.728723 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/443ec657-a3c4-4787-a392-208a02028d88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.728738 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjpf2\" (UniqueName: \"kubernetes.io/projected/443ec657-a3c4-4787-a392-208a02028d88-kube-api-access-wjpf2\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.737475 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:52 crc kubenswrapper[4867]: I1212 08:23:52.737457 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.366552 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"443ec657-a3c4-4787-a392-208a02028d88","Type":"ContainerDied","Data":"54052b5653e046606e6895d1d62f4f9e8fab534ed29915a658951347a42b7b72"} Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.366637 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.366805 4867 scope.go:117] "RemoveContainer" containerID="8bd764366adb88f964f492d480139f012ebb8ac8810e901bfb7238dd7f77da03" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.395728 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.408216 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.421571 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:53 crc kubenswrapper[4867]: E1212 08:23:53.422057 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="443ec657-a3c4-4787-a392-208a02028d88" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.422075 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="443ec657-a3c4-4787-a392-208a02028d88" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.422379 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="443ec657-a3c4-4787-a392-208a02028d88" containerName="nova-cell1-novncproxy-novncproxy" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.423204 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.425528 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.425675 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.426553 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.433077 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.542021 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghzd\" (UniqueName: \"kubernetes.io/projected/4700bc38-07a1-41ea-984c-a51b6db55929-kube-api-access-kghzd\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.542116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.542188 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.542250 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.542434 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.644768 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.645160 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghzd\" (UniqueName: \"kubernetes.io/projected/4700bc38-07a1-41ea-984c-a51b6db55929-kube-api-access-kghzd\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.645241 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.645268 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.645291 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.664292 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.664298 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.664392 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.664519 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4700bc38-07a1-41ea-984c-a51b6db55929-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.666395 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghzd\" (UniqueName: \"kubernetes.io/projected/4700bc38-07a1-41ea-984c-a51b6db55929-kube-api-access-kghzd\") pod \"nova-cell1-novncproxy-0\" (UID: \"4700bc38-07a1-41ea-984c-a51b6db55929\") " pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:53 crc kubenswrapper[4867]: I1212 08:23:53.750916 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.172075 4867 scope.go:117] "RemoveContainer" containerID="dafc4a0d2c9fede68c1bd03410af69ccdd4ebe7141380fcf13a84f35f43a2244" Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.219126 4867 scope.go:117] "RemoveContainer" containerID="2b6e73bd3ad851f8f4e007959a4e7d88246b829788dde949727ae804af2d1de7" Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.263489 4867 scope.go:117] "RemoveContainer" containerID="3e7bd986c549be106a8d6ee6bd71df5784a614dba90580d4ad26b0795ab4e217" Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.295813 4867 scope.go:117] "RemoveContainer" containerID="83f0eabf639e422848a12e485655f4d6f22a4fb3304c77654d9ca0c6d4d445d8" Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.317949 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.380926 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4700bc38-07a1-41ea-984c-a51b6db55929","Type":"ContainerStarted","Data":"9b30b1c46ee8def7b5a700191cb81d7e183fc6a62145f90b86d221de38a16af1"} Dec 12 08:23:54 crc kubenswrapper[4867]: I1212 08:23:54.848495 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="443ec657-a3c4-4787-a392-208a02028d88" path="/var/lib/kubelet/pods/443ec657-a3c4-4787-a392-208a02028d88/volumes" Dec 12 08:23:55 crc kubenswrapper[4867]: I1212 08:23:55.394565 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4700bc38-07a1-41ea-984c-a51b6db55929","Type":"ContainerStarted","Data":"5ccf007489e9bb6494ebbf7394e78111901037d209fda42064be2c845b082d0a"} Dec 12 08:23:55 crc kubenswrapper[4867]: I1212 08:23:55.416977 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.416955964 podStartE2EDuration="2.416955964s" podCreationTimestamp="2025-12-12 08:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:55.415506828 +0000 UTC m=+5722.986888117" watchObservedRunningTime="2025-12-12 08:23:55.416955964 +0000 UTC m=+5722.988337283" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.424547 4867 generic.go:334] "Generic (PLEG): container finished" podID="eebaa5be-47cc-43fa-aba6-713c06a7df46" containerID="7f841a1b3956dd375cc5b9c116e73d4359660288330f8b616f58f4257e61a561" exitCode=137 Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.425478 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eebaa5be-47cc-43fa-aba6-713c06a7df46","Type":"ContainerDied","Data":"7f841a1b3956dd375cc5b9c116e73d4359660288330f8b616f58f4257e61a561"} Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.747096 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.828988 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data\") pod \"eebaa5be-47cc-43fa-aba6-713c06a7df46\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.829146 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle\") pod \"eebaa5be-47cc-43fa-aba6-713c06a7df46\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.829175 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtxwk\" (UniqueName: \"kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk\") pod \"eebaa5be-47cc-43fa-aba6-713c06a7df46\" (UID: \"eebaa5be-47cc-43fa-aba6-713c06a7df46\") " Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.834644 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk" (OuterVolumeSpecName: "kube-api-access-mtxwk") pod "eebaa5be-47cc-43fa-aba6-713c06a7df46" (UID: "eebaa5be-47cc-43fa-aba6-713c06a7df46"). InnerVolumeSpecName "kube-api-access-mtxwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.857737 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eebaa5be-47cc-43fa-aba6-713c06a7df46" (UID: "eebaa5be-47cc-43fa-aba6-713c06a7df46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.858921 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data" (OuterVolumeSpecName: "config-data") pod "eebaa5be-47cc-43fa-aba6-713c06a7df46" (UID: "eebaa5be-47cc-43fa-aba6-713c06a7df46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.931544 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.931586 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtxwk\" (UniqueName: \"kubernetes.io/projected/eebaa5be-47cc-43fa-aba6-713c06a7df46-kube-api-access-mtxwk\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:56 crc kubenswrapper[4867]: I1212 08:23:56.931600 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebaa5be-47cc-43fa-aba6-713c06a7df46-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.435684 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eebaa5be-47cc-43fa-aba6-713c06a7df46","Type":"ContainerDied","Data":"38d40b0bfe81ce3e734a02e262e1c6e0710717c4d1259ebaf100a8839546fda7"} Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.435769 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.436186 4867 scope.go:117] "RemoveContainer" containerID="7f841a1b3956dd375cc5b9c116e73d4359660288330f8b616f58f4257e61a561" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.487496 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.506951 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.525685 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:57 crc kubenswrapper[4867]: E1212 08:23:57.526212 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebaa5be-47cc-43fa-aba6-713c06a7df46" containerName="nova-scheduler-scheduler" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.526251 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebaa5be-47cc-43fa-aba6-713c06a7df46" containerName="nova-scheduler-scheduler" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.526497 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebaa5be-47cc-43fa-aba6-713c06a7df46" containerName="nova-scheduler-scheduler" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.527364 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.529293 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.547011 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.644735 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.644886 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.644945 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wjfq\" (UniqueName: \"kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.746078 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.746151 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wjfq\" (UniqueName: \"kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.746237 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.750994 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.753051 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.763000 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wjfq\" (UniqueName: \"kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq\") pod \"nova-scheduler-0\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " pod="openstack/nova-scheduler-0" Dec 12 08:23:57 crc kubenswrapper[4867]: I1212 08:23:57.850607 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.284806 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:23:58 crc kubenswrapper[4867]: W1212 08:23:58.289510 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18e0b1c1_4643_42d8_957f_5d2ce38f5e64.slice/crio-be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125 WatchSource:0}: Error finding container be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125: Status 404 returned error can't find the container with id be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125 Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.455641 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18e0b1c1-4643-42d8-957f-5d2ce38f5e64","Type":"ContainerStarted","Data":"be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125"} Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.753757 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.848959 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eebaa5be-47cc-43fa-aba6-713c06a7df46" path="/var/lib/kubelet/pods/eebaa5be-47cc-43fa-aba6-713c06a7df46/volumes" Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.989333 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:23:58 crc kubenswrapper[4867]: I1212 08:23:58.989402 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:23:59 crc kubenswrapper[4867]: I1212 08:23:59.475645 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18e0b1c1-4643-42d8-957f-5d2ce38f5e64","Type":"ContainerStarted","Data":"3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2"} Dec 12 08:24:00 crc kubenswrapper[4867]: I1212 08:24:00.522376 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:00 crc kubenswrapper[4867]: I1212 08:24:00.522395 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:02 crc kubenswrapper[4867]: I1212 08:24:02.737445 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:02 crc kubenswrapper[4867]: I1212 08:24:02.737493 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:02 crc kubenswrapper[4867]: I1212 08:24:02.851126 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 08:24:03 crc kubenswrapper[4867]: I1212 08:24:03.752431 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:24:03 crc kubenswrapper[4867]: I1212 08:24:03.773633 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:24:03 crc kubenswrapper[4867]: I1212 08:24:03.793175 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=6.793152537 podStartE2EDuration="6.793152537s" podCreationTimestamp="2025-12-12 08:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:23:59.493272898 +0000 UTC m=+5727.064654167" watchObservedRunningTime="2025-12-12 08:24:03.793152537 +0000 UTC m=+5731.364533806" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.531027 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.701166 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-c48nw"] Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.702329 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.705543 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.706155 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.717191 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c48nw"] Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.789582 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm69w\" (UniqueName: \"kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.789672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.789719 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.789797 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.891953 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.892145 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm69w\" (UniqueName: \"kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.892188 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.892239 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.898062 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.900728 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.905941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:04 crc kubenswrapper[4867]: I1212 08:24:04.926093 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm69w\" (UniqueName: \"kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w\") pod \"nova-cell1-cell-mapping-c48nw\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:05 crc kubenswrapper[4867]: I1212 08:24:05.029247 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:05 crc kubenswrapper[4867]: I1212 08:24:05.488409 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-c48nw"] Dec 12 08:24:05 crc kubenswrapper[4867]: W1212 08:24:05.495613 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod620fe78b_77e2_4dfa_bae8_a56a029cf5e7.slice/crio-af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681 WatchSource:0}: Error finding container af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681: Status 404 returned error can't find the container with id af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681 Dec 12 08:24:05 crc kubenswrapper[4867]: I1212 08:24:05.522942 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c48nw" event={"ID":"620fe78b-77e2-4dfa-bae8-a56a029cf5e7","Type":"ContainerStarted","Data":"af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681"} Dec 12 08:24:06 crc kubenswrapper[4867]: I1212 08:24:06.532031 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c48nw" event={"ID":"620fe78b-77e2-4dfa-bae8-a56a029cf5e7","Type":"ContainerStarted","Data":"6dd9748445cd52bcb74b1787d737bc9d2468ca63ddaf923d21c0e72c273f6c6f"} Dec 12 08:24:06 crc kubenswrapper[4867]: I1212 08:24:06.548835 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-c48nw" podStartSLOduration=2.548818035 podStartE2EDuration="2.548818035s" podCreationTimestamp="2025-12-12 08:24:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:24:06.54861341 +0000 UTC m=+5734.119994679" watchObservedRunningTime="2025-12-12 08:24:06.548818035 +0000 UTC m=+5734.120199294" Dec 12 08:24:07 crc kubenswrapper[4867]: I1212 08:24:07.852010 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 08:24:07 crc kubenswrapper[4867]: I1212 08:24:07.922192 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 08:24:08 crc kubenswrapper[4867]: I1212 08:24:08.588001 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 08:24:10 crc kubenswrapper[4867]: I1212 08:24:10.522422 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:10 crc kubenswrapper[4867]: I1212 08:24:10.522566 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.85:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:11 crc kubenswrapper[4867]: I1212 08:24:11.590265 4867 generic.go:334] "Generic (PLEG): container finished" podID="620fe78b-77e2-4dfa-bae8-a56a029cf5e7" containerID="6dd9748445cd52bcb74b1787d737bc9d2468ca63ddaf923d21c0e72c273f6c6f" exitCode=0 Dec 12 08:24:11 crc kubenswrapper[4867]: I1212 08:24:11.590321 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c48nw" event={"ID":"620fe78b-77e2-4dfa-bae8-a56a029cf5e7","Type":"ContainerDied","Data":"6dd9748445cd52bcb74b1787d737bc9d2468ca63ddaf923d21c0e72c273f6c6f"} Dec 12 08:24:11 crc kubenswrapper[4867]: I1212 08:24:11.654573 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:24:11 crc kubenswrapper[4867]: I1212 08:24:11.654644 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:24:12 crc kubenswrapper[4867]: I1212 08:24:12.736466 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:12 crc kubenswrapper[4867]: I1212 08:24:12.736480 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.86:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.000731 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.070855 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm69w\" (UniqueName: \"kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w\") pod \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.071005 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts\") pod \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.071053 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data\") pod \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.071274 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle\") pod \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\" (UID: \"620fe78b-77e2-4dfa-bae8-a56a029cf5e7\") " Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.080597 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w" (OuterVolumeSpecName: "kube-api-access-rm69w") pod "620fe78b-77e2-4dfa-bae8-a56a029cf5e7" (UID: "620fe78b-77e2-4dfa-bae8-a56a029cf5e7"). InnerVolumeSpecName "kube-api-access-rm69w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.093475 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts" (OuterVolumeSpecName: "scripts") pod "620fe78b-77e2-4dfa-bae8-a56a029cf5e7" (UID: "620fe78b-77e2-4dfa-bae8-a56a029cf5e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.148488 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "620fe78b-77e2-4dfa-bae8-a56a029cf5e7" (UID: "620fe78b-77e2-4dfa-bae8-a56a029cf5e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.172929 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.172977 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm69w\" (UniqueName: \"kubernetes.io/projected/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-kube-api-access-rm69w\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.172989 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.174965 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data" (OuterVolumeSpecName: "config-data") pod "620fe78b-77e2-4dfa-bae8-a56a029cf5e7" (UID: "620fe78b-77e2-4dfa-bae8-a56a029cf5e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.274467 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620fe78b-77e2-4dfa-bae8-a56a029cf5e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.612262 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-c48nw" event={"ID":"620fe78b-77e2-4dfa-bae8-a56a029cf5e7","Type":"ContainerDied","Data":"af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681"} Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.612405 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af0389547c1c08790fdb7129581936377cc4e5bc3ee8f5827d04e5fcfae2f681" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.612352 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-c48nw" Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.790068 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.790437 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" containerID="cri-o://7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9" gracePeriod=30 Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.790535 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" containerID="cri-o://a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1" gracePeriod=30 Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.802165 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.802414 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" containerID="cri-o://3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" gracePeriod=30 Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.835984 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.836265 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" containerID="cri-o://caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2" gracePeriod=30 Dec 12 08:24:13 crc kubenswrapper[4867]: I1212 08:24:13.836400 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" containerID="cri-o://b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9" gracePeriod=30 Dec 12 08:24:14 crc kubenswrapper[4867]: I1212 08:24:14.623119 4867 generic.go:334] "Generic (PLEG): container finished" podID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerID="7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9" exitCode=143 Dec 12 08:24:14 crc kubenswrapper[4867]: I1212 08:24:14.623481 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerDied","Data":"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9"} Dec 12 08:24:14 crc kubenswrapper[4867]: I1212 08:24:14.625765 4867 generic.go:334] "Generic (PLEG): container finished" podID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerID="caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2" exitCode=143 Dec 12 08:24:14 crc kubenswrapper[4867]: I1212 08:24:14.625789 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerDied","Data":"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2"} Dec 12 08:24:17 crc kubenswrapper[4867]: E1212 08:24:17.852950 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:17 crc kubenswrapper[4867]: E1212 08:24:17.855988 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:17 crc kubenswrapper[4867]: E1212 08:24:17.858290 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:17 crc kubenswrapper[4867]: E1212 08:24:17.858426 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:22 crc kubenswrapper[4867]: E1212 08:24:22.852888 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:22 crc kubenswrapper[4867]: E1212 08:24:22.854918 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:22 crc kubenswrapper[4867]: E1212 08:24:22.856427 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:22 crc kubenswrapper[4867]: E1212 08:24:22.856478 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.636646 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.642669 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.732183 4867 generic.go:334] "Generic (PLEG): container finished" podID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerID="a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1" exitCode=0 Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.732268 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.732264 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerDied","Data":"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1"} Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.732652 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868","Type":"ContainerDied","Data":"f25d2203a5d6d4c20face71539694536414624b04119b34fac5ac193992afff3"} Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.732681 4867 scope.go:117] "RemoveContainer" containerID="a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.734515 4867 generic.go:334] "Generic (PLEG): container finished" podID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerID="b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9" exitCode=0 Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.734542 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerDied","Data":"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9"} Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.734560 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"70bbcd36-ede6-46af-a696-1c501665a3d3","Type":"ContainerDied","Data":"401b82efcba1189838bf526be7a3382491d14f7625ff3db5b76cd4e9b3987366"} Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.734651 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.753716 4867 scope.go:117] "RemoveContainer" containerID="7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.770541 4867 scope.go:117] "RemoveContainer" containerID="a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1" Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.771063 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1\": container with ID starting with a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1 not found: ID does not exist" containerID="a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.771107 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1"} err="failed to get container status \"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1\": rpc error: code = NotFound desc = could not find container \"a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1\": container with ID starting with a925ad893a52d2385d4010211d7e5e626c22c65b7071d9318fbd8efdf885e6e1 not found: ID does not exist" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.771133 4867 scope.go:117] "RemoveContainer" containerID="7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9" Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.771494 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9\": container with ID starting with 7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9 not found: ID does not exist" containerID="7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.771523 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9"} err="failed to get container status \"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9\": rpc error: code = NotFound desc = could not find container \"7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9\": container with ID starting with 7473f8043fc267f446f4fa3535efb820875ad75b2fe960b866276122ab320ea9 not found: ID does not exist" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.771536 4867 scope.go:117] "RemoveContainer" containerID="b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.788116 4867 scope.go:117] "RemoveContainer" containerID="caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.805716 4867 scope.go:117] "RemoveContainer" containerID="b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9" Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.806208 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9\": container with ID starting with b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9 not found: ID does not exist" containerID="b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.806337 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9"} err="failed to get container status \"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9\": rpc error: code = NotFound desc = could not find container \"b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9\": container with ID starting with b0bfef3394bb14ab5b80512458357e2a99faf738119e594de3fb75f46f77c5d9 not found: ID does not exist" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.806367 4867 scope.go:117] "RemoveContainer" containerID="caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2" Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.806858 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2\": container with ID starting with caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2 not found: ID does not exist" containerID="caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.806911 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2"} err="failed to get container status \"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2\": rpc error: code = NotFound desc = could not find container \"caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2\": container with ID starting with caeca37f37146d9b8bd8dc8cd4f225eca8efbc9b0cd81ef1c5f4abe77406d9d2 not found: ID does not exist" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.833890 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hdtv\" (UniqueName: \"kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv\") pod \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834103 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data\") pod \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834151 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs\") pod \"70bbcd36-ede6-46af-a696-1c501665a3d3\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834189 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs\") pod \"70bbcd36-ede6-46af-a696-1c501665a3d3\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834249 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data\") pod \"70bbcd36-ede6-46af-a696-1c501665a3d3\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834294 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle\") pod \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834346 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs\") pod \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\" (UID: \"2f570cb5-4ab3-4ed3-b7e9-59eebfa27868\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle\") pod \"70bbcd36-ede6-46af-a696-1c501665a3d3\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.834424 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk5rq\" (UniqueName: \"kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq\") pod \"70bbcd36-ede6-46af-a696-1c501665a3d3\" (UID: \"70bbcd36-ede6-46af-a696-1c501665a3d3\") " Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.836072 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs" (OuterVolumeSpecName: "logs") pod "70bbcd36-ede6-46af-a696-1c501665a3d3" (UID: "70bbcd36-ede6-46af-a696-1c501665a3d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.836158 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs" (OuterVolumeSpecName: "logs") pod "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" (UID: "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.841506 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq" (OuterVolumeSpecName: "kube-api-access-bk5rq") pod "70bbcd36-ede6-46af-a696-1c501665a3d3" (UID: "70bbcd36-ede6-46af-a696-1c501665a3d3"). InnerVolumeSpecName "kube-api-access-bk5rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.841572 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv" (OuterVolumeSpecName: "kube-api-access-2hdtv") pod "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" (UID: "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868"). InnerVolumeSpecName "kube-api-access-2hdtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.852783 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.854706 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.855868 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:27 crc kubenswrapper[4867]: E1212 08:24:27.855908 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.862453 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" (UID: "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.865385 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data" (OuterVolumeSpecName: "config-data") pod "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" (UID: "2f570cb5-4ab3-4ed3-b7e9-59eebfa27868"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.866470 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data" (OuterVolumeSpecName: "config-data") pod "70bbcd36-ede6-46af-a696-1c501665a3d3" (UID: "70bbcd36-ede6-46af-a696-1c501665a3d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.869741 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70bbcd36-ede6-46af-a696-1c501665a3d3" (UID: "70bbcd36-ede6-46af-a696-1c501665a3d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.883374 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "70bbcd36-ede6-46af-a696-1c501665a3d3" (UID: "70bbcd36-ede6-46af-a696-1c501665a3d3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936079 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hdtv\" (UniqueName: \"kubernetes.io/projected/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-kube-api-access-2hdtv\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936116 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936127 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936136 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70bbcd36-ede6-46af-a696-1c501665a3d3-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936145 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936153 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936163 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936171 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70bbcd36-ede6-46af-a696-1c501665a3d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:27 crc kubenswrapper[4867]: I1212 08:24:27.936181 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk5rq\" (UniqueName: \"kubernetes.io/projected/70bbcd36-ede6-46af-a696-1c501665a3d3-kube-api-access-bk5rq\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.073982 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.085500 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.097987 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.116692 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133098 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: E1212 08:24:28.133487 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133506 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" Dec 12 08:24:28 crc kubenswrapper[4867]: E1212 08:24:28.133519 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620fe78b-77e2-4dfa-bae8-a56a029cf5e7" containerName="nova-manage" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133525 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="620fe78b-77e2-4dfa-bae8-a56a029cf5e7" containerName="nova-manage" Dec 12 08:24:28 crc kubenswrapper[4867]: E1212 08:24:28.133538 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133546 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" Dec 12 08:24:28 crc kubenswrapper[4867]: E1212 08:24:28.133566 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133573 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" Dec 12 08:24:28 crc kubenswrapper[4867]: E1212 08:24:28.133596 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133603 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133808 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-metadata" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133821 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-log" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133831 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="620fe78b-77e2-4dfa-bae8-a56a029cf5e7" containerName="nova-manage" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133841 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" containerName="nova-metadata-log" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.133857 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" containerName="nova-api-api" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.134930 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.136641 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.151632 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.153266 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.157368 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.157379 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.167148 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.182367 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.241464 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z87tv\" (UniqueName: \"kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.241644 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.241719 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.241784 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344100 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z87tv\" (UniqueName: \"kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344272 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344370 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344519 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344704 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.344890 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.345062 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.345144 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.345150 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5mb7\" (UniqueName: \"kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.348455 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.349033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.358904 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z87tv\" (UniqueName: \"kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv\") pod \"nova-api-0\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.447082 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.447188 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.447239 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5mb7\" (UniqueName: \"kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.447316 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.447709 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.448290 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.450928 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.451132 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.451846 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.454827 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.465563 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5mb7\" (UniqueName: \"kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7\") pod \"nova-metadata-0\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.471722 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.848710 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f570cb5-4ab3-4ed3-b7e9-59eebfa27868" path="/var/lib/kubelet/pods/2f570cb5-4ab3-4ed3-b7e9-59eebfa27868/volumes" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.850155 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70bbcd36-ede6-46af-a696-1c501665a3d3" path="/var/lib/kubelet/pods/70bbcd36-ede6-46af-a696-1c501665a3d3/volumes" Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.989049 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:24:28 crc kubenswrapper[4867]: I1212 08:24:28.989101 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:28.999968 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.062831 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.773596 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerStarted","Data":"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.773974 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerStarted","Data":"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.773990 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerStarted","Data":"d9f4bff2bf9d039256ffd8edfe241b2f74f5a535dbaa500a3ecab45e955cdba4"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.778696 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerStarted","Data":"d649cf88d7b70e4ebbf21e46f35b225baed9d1ff65d86e0a97bfeb0016203480"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.778746 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerStarted","Data":"6a6a3fad860bb48f1cadf4a3458b5e082f4e5f5c2b4a2f11003c46ac6186158e"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.778760 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerStarted","Data":"1ba42ea77700ac3ff526008cc7af34fbf4eb273115c0f27be6499228d69db9a9"} Dec 12 08:24:29 crc kubenswrapper[4867]: I1212 08:24:29.800805 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.800786251 podStartE2EDuration="1.800786251s" podCreationTimestamp="2025-12-12 08:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:24:29.791023118 +0000 UTC m=+5757.362404387" watchObservedRunningTime="2025-12-12 08:24:29.800786251 +0000 UTC m=+5757.372167660" Dec 12 08:24:32 crc kubenswrapper[4867]: E1212 08:24:32.853252 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:32 crc kubenswrapper[4867]: E1212 08:24:32.855708 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:32 crc kubenswrapper[4867]: E1212 08:24:32.857152 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:32 crc kubenswrapper[4867]: E1212 08:24:32.857205 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:33 crc kubenswrapper[4867]: I1212 08:24:33.472098 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 08:24:33 crc kubenswrapper[4867]: I1212 08:24:33.472178 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 08:24:37 crc kubenswrapper[4867]: E1212 08:24:37.853271 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:37 crc kubenswrapper[4867]: E1212 08:24:37.855595 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:37 crc kubenswrapper[4867]: E1212 08:24:37.857044 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:37 crc kubenswrapper[4867]: E1212 08:24:37.857097 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:38 crc kubenswrapper[4867]: I1212 08:24:38.455544 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:24:38 crc kubenswrapper[4867]: I1212 08:24:38.455602 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:24:38 crc kubenswrapper[4867]: I1212 08:24:38.472148 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 08:24:38 crc kubenswrapper[4867]: I1212 08:24:38.472261 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 08:24:39 crc kubenswrapper[4867]: I1212 08:24:39.497526 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:39 crc kubenswrapper[4867]: I1212 08:24:39.550448 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:39 crc kubenswrapper[4867]: I1212 08:24:39.550448 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:39 crc kubenswrapper[4867]: I1212 08:24:39.550464 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:24:42 crc kubenswrapper[4867]: E1212 08:24:42.852504 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:42 crc kubenswrapper[4867]: E1212 08:24:42.854649 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:42 crc kubenswrapper[4867]: E1212 08:24:42.856097 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 12 08:24:42 crc kubenswrapper[4867]: E1212 08:24:42.856158 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:43 crc kubenswrapper[4867]: W1212 08:24:43.840371 4867 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18e0b1c1_4643_42d8_957f_5d2ce38f5e64.slice/crio-be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125/cpu.weight": read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18e0b1c1_4643_42d8_957f_5d2ce38f5e64.slice/crio-be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125/cpu.weight: no such device Dec 12 08:24:43 crc kubenswrapper[4867]: I1212 08:24:43.906342 4867 generic.go:334] "Generic (PLEG): container finished" podID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" exitCode=137 Dec 12 08:24:43 crc kubenswrapper[4867]: I1212 08:24:43.906384 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18e0b1c1-4643-42d8-957f-5d2ce38f5e64","Type":"ContainerDied","Data":"3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2"} Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.207194 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.228817 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=16.228797267 podStartE2EDuration="16.228797267s" podCreationTimestamp="2025-12-12 08:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:24:29.811769593 +0000 UTC m=+5757.383150862" watchObservedRunningTime="2025-12-12 08:24:44.228797267 +0000 UTC m=+5771.800178536" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.343121 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data\") pod \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.343262 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle\") pod \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.343299 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wjfq\" (UniqueName: \"kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq\") pod \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\" (UID: \"18e0b1c1-4643-42d8-957f-5d2ce38f5e64\") " Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.349059 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq" (OuterVolumeSpecName: "kube-api-access-7wjfq") pod "18e0b1c1-4643-42d8-957f-5d2ce38f5e64" (UID: "18e0b1c1-4643-42d8-957f-5d2ce38f5e64"). InnerVolumeSpecName "kube-api-access-7wjfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.374059 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data" (OuterVolumeSpecName: "config-data") pod "18e0b1c1-4643-42d8-957f-5d2ce38f5e64" (UID: "18e0b1c1-4643-42d8-957f-5d2ce38f5e64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.378471 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18e0b1c1-4643-42d8-957f-5d2ce38f5e64" (UID: "18e0b1c1-4643-42d8-957f-5d2ce38f5e64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.445527 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.445574 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wjfq\" (UniqueName: \"kubernetes.io/projected/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-kube-api-access-7wjfq\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.445588 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18e0b1c1-4643-42d8-957f-5d2ce38f5e64-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.916585 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"18e0b1c1-4643-42d8-957f-5d2ce38f5e64","Type":"ContainerDied","Data":"be1ea7d352cf963ee61ef90c5bfe348897b18bbc76d643a8feb5dd229f177125"} Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.916666 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.916890 4867 scope.go:117] "RemoveContainer" containerID="3163062ae7b98135273afaa419f8d57bdf9dd69ddc49f1ee83b924e1f10468a2" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.941954 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.954126 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.976809 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:44 crc kubenswrapper[4867]: E1212 08:24:44.977563 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.977584 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.977746 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" containerName="nova-scheduler-scheduler" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.978386 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.983094 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 08:24:44 crc kubenswrapper[4867]: I1212 08:24:44.993268 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.157710 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.157816 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.158036 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9fxp\" (UniqueName: \"kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.259623 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.259689 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9fxp\" (UniqueName: \"kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.259801 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.269989 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.270598 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.283337 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9fxp\" (UniqueName: \"kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp\") pod \"nova-scheduler-0\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.294477 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.898563 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 08:24:45 crc kubenswrapper[4867]: I1212 08:24:45.929591 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"750d05ab-0045-4a18-9fe8-e9d59566728c","Type":"ContainerStarted","Data":"fff8d2d2e8ae39fbd011330ccaa025416fe47ee4e13502ae16581a45c85bfa8f"} Dec 12 08:24:46 crc kubenswrapper[4867]: I1212 08:24:46.848750 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18e0b1c1-4643-42d8-957f-5d2ce38f5e64" path="/var/lib/kubelet/pods/18e0b1c1-4643-42d8-957f-5d2ce38f5e64/volumes" Dec 12 08:24:46 crc kubenswrapper[4867]: I1212 08:24:46.940135 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"750d05ab-0045-4a18-9fe8-e9d59566728c","Type":"ContainerStarted","Data":"94536244d42ee782cd783729152d4af2480ac426bb93d508ffffe5e952fb91ac"} Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.459190 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.459871 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.460991 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.462757 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.477399 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.485859 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.485832114 podStartE2EDuration="4.485832114s" podCreationTimestamp="2025-12-12 08:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:24:46.960472211 +0000 UTC m=+5774.531853490" watchObservedRunningTime="2025-12-12 08:24:48.485832114 +0000 UTC m=+5776.057213383" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.492690 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.500083 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.501664 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.964669 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:24:48 crc kubenswrapper[4867]: I1212 08:24:48.967774 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.143922 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.146146 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.164155 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.250099 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.250289 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.250426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.250556 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qffx\" (UniqueName: \"kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.250657 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.353047 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qffx\" (UniqueName: \"kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.353147 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.353253 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.353343 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.353404 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.354098 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.354144 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.354272 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.354357 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.375733 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qffx\" (UniqueName: \"kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx\") pod \"dnsmasq-dns-6bd7f7d77f-dm22f\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.471890 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.922564 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:24:49 crc kubenswrapper[4867]: I1212 08:24:49.976852 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" event={"ID":"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9","Type":"ContainerStarted","Data":"923130bb40ad8626bbb78e02c3b37c6afdc0602d1b514d98ddbbfcb9ebbd3563"} Dec 12 08:24:50 crc kubenswrapper[4867]: I1212 08:24:50.295320 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 08:24:50 crc kubenswrapper[4867]: I1212 08:24:50.985511 4867 generic.go:334] "Generic (PLEG): container finished" podID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerID="072c8049ad88ff4bf15d8cc3db9a8a70266ebaf79ffd281fa871d1b740e79d98" exitCode=0 Dec 12 08:24:50 crc kubenswrapper[4867]: I1212 08:24:50.987215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" event={"ID":"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9","Type":"ContainerDied","Data":"072c8049ad88ff4bf15d8cc3db9a8a70266ebaf79ffd281fa871d1b740e79d98"} Dec 12 08:24:51 crc kubenswrapper[4867]: I1212 08:24:51.673152 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:51 crc kubenswrapper[4867]: I1212 08:24:51.998736 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" event={"ID":"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9","Type":"ContainerStarted","Data":"a5d29a5ea3cfb4fc1f1d752ab310eaaf6d1a1c51b97aafb0d88df751920034c0"} Dec 12 08:24:51 crc kubenswrapper[4867]: I1212 08:24:51.998888 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-log" containerID="cri-o://e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a" gracePeriod=30 Dec 12 08:24:51 crc kubenswrapper[4867]: I1212 08:24:51.998928 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-api" containerID="cri-o://0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7" gracePeriod=30 Dec 12 08:24:52 crc kubenswrapper[4867]: I1212 08:24:52.029197 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" podStartSLOduration=3.029175459 podStartE2EDuration="3.029175459s" podCreationTimestamp="2025-12-12 08:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:24:52.028852461 +0000 UTC m=+5779.600233740" watchObservedRunningTime="2025-12-12 08:24:52.029175459 +0000 UTC m=+5779.600556728" Dec 12 08:24:53 crc kubenswrapper[4867]: I1212 08:24:53.006473 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:54 crc kubenswrapper[4867]: I1212 08:24:54.038840 4867 generic.go:334] "Generic (PLEG): container finished" podID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerID="e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a" exitCode=143 Dec 12 08:24:54 crc kubenswrapper[4867]: I1212 08:24:54.039035 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerDied","Data":"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a"} Dec 12 08:24:55 crc kubenswrapper[4867]: I1212 08:24:55.295930 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 08:24:55 crc kubenswrapper[4867]: I1212 08:24:55.324820 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 08:24:56 crc kubenswrapper[4867]: I1212 08:24:56.082978 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 08:24:57 crc kubenswrapper[4867]: I1212 08:24:57.908667 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.026652 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle\") pod \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.026732 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data\") pod \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.026825 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z87tv\" (UniqueName: \"kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv\") pod \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.026858 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs\") pod \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\" (UID: \"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63\") " Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.027828 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs" (OuterVolumeSpecName: "logs") pod "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" (UID: "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.033906 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv" (OuterVolumeSpecName: "kube-api-access-z87tv") pod "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" (UID: "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63"). InnerVolumeSpecName "kube-api-access-z87tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.057496 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" (UID: "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.080175 4867 generic.go:334] "Generic (PLEG): container finished" podID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerID="0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7" exitCode=0 Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.080222 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerDied","Data":"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7"} Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.080274 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1aefbb7b-5188-4bf1-a5f2-84b12bb76a63","Type":"ContainerDied","Data":"d9f4bff2bf9d039256ffd8edfe241b2f74f5a535dbaa500a3ecab45e955cdba4"} Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.080290 4867 scope.go:117] "RemoveContainer" containerID="0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.080612 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.084360 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data" (OuterVolumeSpecName: "config-data") pod "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" (UID: "1aefbb7b-5188-4bf1-a5f2-84b12bb76a63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.131193 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.131257 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z87tv\" (UniqueName: \"kubernetes.io/projected/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-kube-api-access-z87tv\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.131268 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.131276 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.161832 4867 scope.go:117] "RemoveContainer" containerID="e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.183403 4867 scope.go:117] "RemoveContainer" containerID="0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7" Dec 12 08:24:58 crc kubenswrapper[4867]: E1212 08:24:58.187728 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7\": container with ID starting with 0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7 not found: ID does not exist" containerID="0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.187784 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7"} err="failed to get container status \"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7\": rpc error: code = NotFound desc = could not find container \"0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7\": container with ID starting with 0aef72714e3b6403f8d12a8453e7e34501851f9229077063a58aa60a25f22fd7 not found: ID does not exist" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.187814 4867 scope.go:117] "RemoveContainer" containerID="e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a" Dec 12 08:24:58 crc kubenswrapper[4867]: E1212 08:24:58.188290 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a\": container with ID starting with e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a not found: ID does not exist" containerID="e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.188339 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a"} err="failed to get container status \"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a\": rpc error: code = NotFound desc = could not find container \"e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a\": container with ID starting with e6cc4d41a80381fdb91cd625e6ef3d6b8c9bf5153ef4ed47614a844a6f54424a not found: ID does not exist" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.417680 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.432128 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.451546 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:58 crc kubenswrapper[4867]: E1212 08:24:58.452331 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-api" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.452360 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-api" Dec 12 08:24:58 crc kubenswrapper[4867]: E1212 08:24:58.452386 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-log" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.452400 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-log" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.452701 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-api" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.452728 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" containerName="nova-api-log" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.454536 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.456293 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.457937 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.458433 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.461798 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539092 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539151 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539279 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539377 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539469 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msgdv\" (UniqueName: \"kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.539565 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641024 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641098 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641127 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641157 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641256 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641331 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msgdv\" (UniqueName: \"kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.641779 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.646803 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.647134 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.647311 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.647518 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.664716 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msgdv\" (UniqueName: \"kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv\") pod \"nova-api-0\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.798184 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.852145 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aefbb7b-5188-4bf1-a5f2-84b12bb76a63" path="/var/lib/kubelet/pods/1aefbb7b-5188-4bf1-a5f2-84b12bb76a63/volumes" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.989257 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.989326 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.989373 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.990086 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:24:58 crc kubenswrapper[4867]: I1212 08:24:58.990141 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52" gracePeriod=600 Dec 12 08:24:59 crc kubenswrapper[4867]: I1212 08:24:59.306218 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 08:24:59 crc kubenswrapper[4867]: W1212 08:24:59.308769 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16243454_7b73_4448_b1c6_089261fef413.slice/crio-aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320 WatchSource:0}: Error finding container aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320: Status 404 returned error can't find the container with id aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320 Dec 12 08:24:59 crc kubenswrapper[4867]: I1212 08:24:59.473410 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:24:59 crc kubenswrapper[4867]: I1212 08:24:59.546297 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:24:59 crc kubenswrapper[4867]: I1212 08:24:59.546604 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="dnsmasq-dns" containerID="cri-o://527e851f4b1a85e49e4d67716cc580ab031c8f6e47294d3515ba32dccc5f94a8" gracePeriod=10 Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.102121 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerStarted","Data":"6d5a5645c2bac9e863274b960647d85e8f37b212933965ce7e5c348ffaac2620"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.102733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerStarted","Data":"3bb1311ff7539471b027757640ae7cc8d243734703689b5318699c5117b8153f"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.102749 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerStarted","Data":"aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.109645 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52" exitCode=0 Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.109716 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.109739 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.109755 4867 scope.go:117] "RemoveContainer" containerID="c7bc782ae77e8cea56a804b095d353370eb8a9c08ecd6ce8cd68d589b4e56570" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.112647 4867 generic.go:334] "Generic (PLEG): container finished" podID="c33f7309-f9a3-4add-9887-749f61676452" containerID="527e851f4b1a85e49e4d67716cc580ab031c8f6e47294d3515ba32dccc5f94a8" exitCode=0 Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.112674 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" event={"ID":"c33f7309-f9a3-4add-9887-749f61676452","Type":"ContainerDied","Data":"527e851f4b1a85e49e4d67716cc580ab031c8f6e47294d3515ba32dccc5f94a8"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.112694 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" event={"ID":"c33f7309-f9a3-4add-9887-749f61676452","Type":"ContainerDied","Data":"16330b28bce0c748748de6f77ed84a14dec178e6a6f1314f499799df60243402"} Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.112708 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16330b28bce0c748748de6f77ed84a14dec178e6a6f1314f499799df60243402" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.132830 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.206659 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.206633533 podStartE2EDuration="2.206633533s" podCreationTimestamp="2025-12-12 08:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:25:00.157050813 +0000 UTC m=+5787.728432083" watchObservedRunningTime="2025-12-12 08:25:00.206633533 +0000 UTC m=+5787.778014802" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.277658 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb\") pod \"c33f7309-f9a3-4add-9887-749f61676452\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.277736 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc\") pod \"c33f7309-f9a3-4add-9887-749f61676452\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.277788 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mpvw\" (UniqueName: \"kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw\") pod \"c33f7309-f9a3-4add-9887-749f61676452\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.277843 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb\") pod \"c33f7309-f9a3-4add-9887-749f61676452\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.277905 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config\") pod \"c33f7309-f9a3-4add-9887-749f61676452\" (UID: \"c33f7309-f9a3-4add-9887-749f61676452\") " Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.286489 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw" (OuterVolumeSpecName: "kube-api-access-2mpvw") pod "c33f7309-f9a3-4add-9887-749f61676452" (UID: "c33f7309-f9a3-4add-9887-749f61676452"). InnerVolumeSpecName "kube-api-access-2mpvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.346817 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c33f7309-f9a3-4add-9887-749f61676452" (UID: "c33f7309-f9a3-4add-9887-749f61676452"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.352658 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c33f7309-f9a3-4add-9887-749f61676452" (UID: "c33f7309-f9a3-4add-9887-749f61676452"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.357361 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config" (OuterVolumeSpecName: "config") pod "c33f7309-f9a3-4add-9887-749f61676452" (UID: "c33f7309-f9a3-4add-9887-749f61676452"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.368997 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c33f7309-f9a3-4add-9887-749f61676452" (UID: "c33f7309-f9a3-4add-9887-749f61676452"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.381052 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.381104 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.381119 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mpvw\" (UniqueName: \"kubernetes.io/projected/c33f7309-f9a3-4add-9887-749f61676452-kube-api-access-2mpvw\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.381132 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:00 crc kubenswrapper[4867]: I1212 08:25:00.381145 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33f7309-f9a3-4add-9887-749f61676452-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:01 crc kubenswrapper[4867]: I1212 08:25:01.129587 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc5b649c9-fx44l" Dec 12 08:25:01 crc kubenswrapper[4867]: I1212 08:25:01.158247 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:25:01 crc kubenswrapper[4867]: I1212 08:25:01.168980 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fc5b649c9-fx44l"] Dec 12 08:25:02 crc kubenswrapper[4867]: I1212 08:25:02.851140 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c33f7309-f9a3-4add-9887-749f61676452" path="/var/lib/kubelet/pods/c33f7309-f9a3-4add-9887-749f61676452/volumes" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.360983 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:05 crc kubenswrapper[4867]: E1212 08:25:05.361914 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="dnsmasq-dns" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.361933 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="dnsmasq-dns" Dec 12 08:25:05 crc kubenswrapper[4867]: E1212 08:25:05.361948 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="init" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.361955 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="init" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.362202 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c33f7309-f9a3-4add-9887-749f61676452" containerName="dnsmasq-dns" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.363981 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.369065 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.483515 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.483580 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.483688 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtbc9\" (UniqueName: \"kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.586272 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.586434 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.586984 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.587048 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.587272 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtbc9\" (UniqueName: \"kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.616641 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtbc9\" (UniqueName: \"kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9\") pod \"certified-operators-5kxkb\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:05 crc kubenswrapper[4867]: I1212 08:25:05.693449 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:06 crc kubenswrapper[4867]: I1212 08:25:06.190858 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:07 crc kubenswrapper[4867]: I1212 08:25:07.187735 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerID="d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493" exitCode=0 Dec 12 08:25:07 crc kubenswrapper[4867]: I1212 08:25:07.187777 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerDied","Data":"d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493"} Dec 12 08:25:07 crc kubenswrapper[4867]: I1212 08:25:07.187997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerStarted","Data":"22830ed783389581b3a694ad9f842c27696e8eb3bc8cae0cd140a83a7949dc87"} Dec 12 08:25:08 crc kubenswrapper[4867]: I1212 08:25:08.219909 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerStarted","Data":"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f"} Dec 12 08:25:08 crc kubenswrapper[4867]: I1212 08:25:08.799161 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:25:08 crc kubenswrapper[4867]: I1212 08:25:08.799551 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 08:25:09 crc kubenswrapper[4867]: I1212 08:25:09.811427 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.94:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:25:09 crc kubenswrapper[4867]: I1212 08:25:09.811431 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.94:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 08:25:10 crc kubenswrapper[4867]: I1212 08:25:10.241031 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerID="46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f" exitCode=0 Dec 12 08:25:10 crc kubenswrapper[4867]: I1212 08:25:10.241077 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerDied","Data":"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f"} Dec 12 08:25:11 crc kubenswrapper[4867]: I1212 08:25:11.252184 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerStarted","Data":"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62"} Dec 12 08:25:11 crc kubenswrapper[4867]: I1212 08:25:11.280786 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5kxkb" podStartSLOduration=2.708156814 podStartE2EDuration="6.280767174s" podCreationTimestamp="2025-12-12 08:25:05 +0000 UTC" firstStartedPulling="2025-12-12 08:25:07.18953713 +0000 UTC m=+5794.760918399" lastFinishedPulling="2025-12-12 08:25:10.76214749 +0000 UTC m=+5798.333528759" observedRunningTime="2025-12-12 08:25:11.274496618 +0000 UTC m=+5798.845877887" watchObservedRunningTime="2025-12-12 08:25:11.280767174 +0000 UTC m=+5798.852148443" Dec 12 08:25:15 crc kubenswrapper[4867]: I1212 08:25:15.694506 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:15 crc kubenswrapper[4867]: I1212 08:25:15.695272 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:15 crc kubenswrapper[4867]: I1212 08:25:15.737504 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:16 crc kubenswrapper[4867]: I1212 08:25:16.338168 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:16 crc kubenswrapper[4867]: I1212 08:25:16.384730 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.306289 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5kxkb" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="registry-server" containerID="cri-o://9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62" gracePeriod=2 Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.791633 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.805165 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.805598 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.805773 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.815677 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.977769 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtbc9\" (UniqueName: \"kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9\") pod \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.978022 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content\") pod \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.978073 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities\") pod \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\" (UID: \"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136\") " Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.978979 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities" (OuterVolumeSpecName: "utilities") pod "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" (UID: "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:18 crc kubenswrapper[4867]: I1212 08:25:18.985119 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9" (OuterVolumeSpecName: "kube-api-access-gtbc9") pod "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" (UID: "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136"). InnerVolumeSpecName "kube-api-access-gtbc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.030315 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" (UID: "f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.081097 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.081432 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.081499 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtbc9\" (UniqueName: \"kubernetes.io/projected/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136-kube-api-access-gtbc9\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.317545 4867 generic.go:334] "Generic (PLEG): container finished" podID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerID="9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62" exitCode=0 Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.317610 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerDied","Data":"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62"} Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.317628 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5kxkb" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.317921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5kxkb" event={"ID":"f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136","Type":"ContainerDied","Data":"22830ed783389581b3a694ad9f842c27696e8eb3bc8cae0cd140a83a7949dc87"} Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.317981 4867 scope.go:117] "RemoveContainer" containerID="9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.318539 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.330571 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.352747 4867 scope.go:117] "RemoveContainer" containerID="46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.386050 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.393376 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5kxkb"] Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.409853 4867 scope.go:117] "RemoveContainer" containerID="d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.450154 4867 scope.go:117] "RemoveContainer" containerID="9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62" Dec 12 08:25:19 crc kubenswrapper[4867]: E1212 08:25:19.450734 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62\": container with ID starting with 9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62 not found: ID does not exist" containerID="9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.450780 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62"} err="failed to get container status \"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62\": rpc error: code = NotFound desc = could not find container \"9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62\": container with ID starting with 9234a0a21179ff94c68184b265eec7e72c5eeb9b078fb08108c6d6d717fbff62 not found: ID does not exist" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.450809 4867 scope.go:117] "RemoveContainer" containerID="46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f" Dec 12 08:25:19 crc kubenswrapper[4867]: E1212 08:25:19.451049 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f\": container with ID starting with 46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f not found: ID does not exist" containerID="46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.451092 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f"} err="failed to get container status \"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f\": rpc error: code = NotFound desc = could not find container \"46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f\": container with ID starting with 46978f77eb9f1c5ab457131936c5849a42cf25ee1bfc41aef8c2fc4c66e8627f not found: ID does not exist" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.451108 4867 scope.go:117] "RemoveContainer" containerID="d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493" Dec 12 08:25:19 crc kubenswrapper[4867]: E1212 08:25:19.451433 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493\": container with ID starting with d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493 not found: ID does not exist" containerID="d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493" Dec 12 08:25:19 crc kubenswrapper[4867]: I1212 08:25:19.451476 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493"} err="failed to get container status \"d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493\": rpc error: code = NotFound desc = could not find container \"d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493\": container with ID starting with d4a01aa09b4b5cb6ef858ca15e57aba460eabbb99f3422993d3cebd015b43493 not found: ID does not exist" Dec 12 08:25:20 crc kubenswrapper[4867]: I1212 08:25:20.850302 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" path="/var/lib/kubelet/pods/f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136/volumes" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.428123 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:25:31 crc kubenswrapper[4867]: E1212 08:25:31.429266 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="registry-server" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.429285 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="registry-server" Dec 12 08:25:31 crc kubenswrapper[4867]: E1212 08:25:31.429342 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="extract-utilities" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.429356 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="extract-utilities" Dec 12 08:25:31 crc kubenswrapper[4867]: E1212 08:25:31.429374 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="extract-content" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.429382 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="extract-content" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.429668 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b945ea-44ef-4c97-ac2a-b9cf8b7ce136" containerName="registry-server" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.431671 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.442167 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-h449n" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.442443 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.442591 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.443322 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.453505 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.498612 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.498861 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-log" containerID="cri-o://d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31" gracePeriod=30 Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.500365 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-httpd" containerID="cri-o://1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370" gracePeriod=30 Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.526254 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.528531 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.598248 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.605563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.605625 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr2xp\" (UniqueName: \"kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.605758 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.605852 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.605949 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.633840 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.634071 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-log" containerID="cri-o://ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c" gracePeriod=30 Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.634486 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-httpd" containerID="cri-o://08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2" gracePeriod=30 Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707739 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707803 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707854 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-564mx\" (UniqueName: \"kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707948 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.707976 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.708044 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.708128 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.708158 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.708188 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr2xp\" (UniqueName: \"kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.710034 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.710456 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.711662 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.728027 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr2xp\" (UniqueName: \"kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.728246 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key\") pod \"horizon-59b4cf6b67-z2g65\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.756764 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.811103 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.811178 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.811198 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.811240 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-564mx\" (UniqueName: \"kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.811281 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.812412 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.812872 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.813097 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.833221 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.846959 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-564mx\" (UniqueName: \"kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx\") pod \"horizon-64cf5f8fb9-d2m4k\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:31 crc kubenswrapper[4867]: I1212 08:25:31.858285 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.449740 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.450309 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerDied","Data":"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c"} Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.450068 4867 generic.go:334] "Generic (PLEG): container finished" podID="f7852810-979e-45c7-b5f4-1f6de231699f" containerID="ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c" exitCode=143 Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.452964 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.454865 4867 generic.go:334] "Generic (PLEG): container finished" podID="e57d947e-37a6-49b7-bdce-cee499065137" containerID="d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31" exitCode=143 Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.454896 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerDied","Data":"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31"} Dec 12 08:25:32 crc kubenswrapper[4867]: I1212 08:25:32.542191 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:25:32 crc kubenswrapper[4867]: W1212 08:25:32.553000 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9493bb26_5213_4a7a_afb9_fa172e800f0c.slice/crio-bdec6494b6f769600e3cb8a0ac6750c3a4474cc743e72ec0983a22e481bd1b1a WatchSource:0}: Error finding container bdec6494b6f769600e3cb8a0ac6750c3a4474cc743e72ec0983a22e481bd1b1a: Status 404 returned error can't find the container with id bdec6494b6f769600e3cb8a0ac6750c3a4474cc743e72ec0983a22e481bd1b1a Dec 12 08:25:33 crc kubenswrapper[4867]: I1212 08:25:33.466072 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerStarted","Data":"dda0b3c36b0c2a30018389074827c1019b09c3a7e38b76e18b99718bbdda553a"} Dec 12 08:25:33 crc kubenswrapper[4867]: I1212 08:25:33.467932 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerStarted","Data":"bdec6494b6f769600e3cb8a0ac6750c3a4474cc743e72ec0983a22e481bd1b1a"} Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.228167 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.272623 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.274461 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.277469 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.287429 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.361508 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.372799 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.372865 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.372896 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.372955 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.372984 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.373027 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsl55\" (UniqueName: \"kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.373092 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.393336 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.394874 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.454104 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480021 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480072 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480097 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480206 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480278 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480380 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsl55\" (UniqueName: \"kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.480551 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.481365 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.481926 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.482890 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.487056 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.496326 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.496586 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.503874 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsl55\" (UniqueName: \"kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55\") pod \"horizon-76dc45d54-sw2qv\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.581891 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.581944 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5sk7\" (UniqueName: \"kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.581980 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.582040 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.582079 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.582100 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.582254 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.610692 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.684150 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.685812 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.685864 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.685910 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.686154 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.686249 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.686282 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5sk7\" (UniqueName: \"kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.686298 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.686327 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.687254 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.690026 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.694985 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.695827 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.710318 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5sk7\" (UniqueName: \"kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7\") pod \"horizon-7d5d9f84d-gcsp7\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:34 crc kubenswrapper[4867]: I1212 08:25:34.719086 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.248693 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.369813 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:25:35 crc kubenswrapper[4867]: W1212 08:25:35.382777 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e86b4f_3c98_4a37_8e2f_866551494bad.slice/crio-b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c WatchSource:0}: Error finding container b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c: Status 404 returned error can't find the container with id b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.418602 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.506204 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerStarted","Data":"b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.508157 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.509333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerStarted","Data":"1bc920922ffd5db408450f400a39510688b651cc9a0570d439e0dbca5e592dc4"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.552616 4867 generic.go:334] "Generic (PLEG): container finished" podID="f7852810-979e-45c7-b5f4-1f6de231699f" containerID="08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2" exitCode=0 Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.552689 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerDied","Data":"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.552718 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7852810-979e-45c7-b5f4-1f6de231699f","Type":"ContainerDied","Data":"fb25bbe31618b12beb9a820c4a8da952fead6812ec870c6560a2f20710b176e9"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.552737 4867 scope.go:117] "RemoveContainer" containerID="08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.552874 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.566498 4867 generic.go:334] "Generic (PLEG): container finished" podID="e57d947e-37a6-49b7-bdce-cee499065137" containerID="1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370" exitCode=0 Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.566549 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerDied","Data":"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.566581 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e57d947e-37a6-49b7-bdce-cee499065137","Type":"ContainerDied","Data":"6e9289aab713b4b7e0d3b05711afeb4725d81a2869a020fd57f012a6dd4176e8"} Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.566716 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.589908 4867 scope.go:117] "RemoveContainer" containerID="ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.617394 4867 scope.go:117] "RemoveContainer" containerID="08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.617843 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2\": container with ID starting with 08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2 not found: ID does not exist" containerID="08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.617884 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2"} err="failed to get container status \"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2\": rpc error: code = NotFound desc = could not find container \"08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2\": container with ID starting with 08e7cbc205b642530a67a88747e2483ec4a38f0fc8890211a3b91ab980f79ff2 not found: ID does not exist" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.617906 4867 scope.go:117] "RemoveContainer" containerID="ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.618333 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c\": container with ID starting with ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c not found: ID does not exist" containerID="ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.618358 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c"} err="failed to get container status \"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c\": rpc error: code = NotFound desc = could not find container \"ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c\": container with ID starting with ac8e42e9d1d7bc0652ad556d819c46c4bd136a58a4056e5f6d1afd1d58bbae8c not found: ID does not exist" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.618371 4867 scope.go:117] "RemoveContainer" containerID="1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.621205 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622169 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622396 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tb2s\" (UniqueName: \"kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622501 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622574 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622627 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622658 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622682 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622707 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622743 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle\") pod \"f7852810-979e-45c7-b5f4-1f6de231699f\" (UID: \"f7852810-979e-45c7-b5f4-1f6de231699f\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622787 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622815 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622838 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.622874 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6b7r\" (UniqueName: \"kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r\") pod \"e57d947e-37a6-49b7-bdce-cee499065137\" (UID: \"e57d947e-37a6-49b7-bdce-cee499065137\") " Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.623177 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.623333 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs" (OuterVolumeSpecName: "logs") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.623478 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.623666 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs" (OuterVolumeSpecName: "logs") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.624249 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.624471 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7852810-979e-45c7-b5f4-1f6de231699f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.624485 4867 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.624497 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57d947e-37a6-49b7-bdce-cee499065137-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.630183 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s" (OuterVolumeSpecName: "kube-api-access-9tb2s") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "kube-api-access-9tb2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.640678 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts" (OuterVolumeSpecName: "scripts") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.640786 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r" (OuterVolumeSpecName: "kube-api-access-n6b7r") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "kube-api-access-n6b7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.666541 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts" (OuterVolumeSpecName: "scripts") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.669142 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.701123 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.702661 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data" (OuterVolumeSpecName: "config-data") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.707356 4867 scope.go:117] "RemoveContainer" containerID="d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.718407 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data" (OuterVolumeSpecName: "config-data") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727047 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727082 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727095 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727106 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727117 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727129 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6b7r\" (UniqueName: \"kubernetes.io/projected/e57d947e-37a6-49b7-bdce-cee499065137-kube-api-access-n6b7r\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727141 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.727375 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tb2s\" (UniqueName: \"kubernetes.io/projected/f7852810-979e-45c7-b5f4-1f6de231699f-kube-api-access-9tb2s\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.739544 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e57d947e-37a6-49b7-bdce-cee499065137" (UID: "e57d947e-37a6-49b7-bdce-cee499065137"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.741020 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f7852810-979e-45c7-b5f4-1f6de231699f" (UID: "f7852810-979e-45c7-b5f4-1f6de231699f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.823692 4867 scope.go:117] "RemoveContainer" containerID="1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.824131 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370\": container with ID starting with 1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370 not found: ID does not exist" containerID="1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.824170 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370"} err="failed to get container status \"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370\": rpc error: code = NotFound desc = could not find container \"1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370\": container with ID starting with 1c4c90fdc5b11bd2c48d3aa9c6f6d0890f6cefe2ccc9f894e5e48b8ffa920370 not found: ID does not exist" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.824195 4867 scope.go:117] "RemoveContainer" containerID="d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.824635 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31\": container with ID starting with d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31 not found: ID does not exist" containerID="d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.824668 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31"} err="failed to get container status \"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31\": rpc error: code = NotFound desc = could not find container \"d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31\": container with ID starting with d07ac9555af1ea1e25b2b20a6c63ca29cc5950f81bf39d2a05292518422a8e31 not found: ID does not exist" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.828572 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7852810-979e-45c7-b5f4-1f6de231699f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.828596 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57d947e-37a6-49b7-bdce-cee499065137-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.918122 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.930277 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.943379 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.958378 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.958862 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.958875 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.958910 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.958917 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.958927 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.958934 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: E1212 08:25:35.958948 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.958953 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.959117 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.959143 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-httpd" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.959157 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57d947e-37a6-49b7-bdce-cee499065137" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.959168 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" containerName="glance-log" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.960119 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.962505 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.962721 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.963114 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.964250 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zqqmt" Dec 12 08:25:35 crc kubenswrapper[4867]: I1212 08:25:35.996084 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.015347 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.023315 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.025370 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.027843 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.027945 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.031972 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033366 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033431 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-logs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033535 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033629 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033651 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkxb\" (UniqueName: \"kubernetes.io/projected/aea8bd9c-4959-457f-b87e-dff93c91f4e7-kube-api-access-9qkxb\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033690 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033706 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhkf\" (UniqueName: \"kubernetes.io/projected/7630d135-1664-41e2-9b9e-76a325f47863-kube-api-access-pvhkf\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033805 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033870 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033910 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.033939 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-logs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.034041 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.034133 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.042825 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.139622 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.139670 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkxb\" (UniqueName: \"kubernetes.io/projected/aea8bd9c-4959-457f-b87e-dff93c91f4e7-kube-api-access-9qkxb\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.139697 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.139717 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhkf\" (UniqueName: \"kubernetes.io/projected/7630d135-1664-41e2-9b9e-76a325f47863-kube-api-access-pvhkf\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.140479 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141206 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141285 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141374 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141408 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-logs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141543 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141635 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141664 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141703 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141783 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-logs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.141877 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.142090 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aea8bd9c-4959-457f-b87e-dff93c91f4e7-logs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.142691 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.146063 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7630d135-1664-41e2-9b9e-76a325f47863-logs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.148165 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.148778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.151981 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.151993 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.155762 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.157623 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aea8bd9c-4959-457f-b87e-dff93c91f4e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.161617 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.162778 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7630d135-1664-41e2-9b9e-76a325f47863-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.165371 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkxb\" (UniqueName: \"kubernetes.io/projected/aea8bd9c-4959-457f-b87e-dff93c91f4e7-kube-api-access-9qkxb\") pod \"glance-default-external-api-0\" (UID: \"aea8bd9c-4959-457f-b87e-dff93c91f4e7\") " pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.165940 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhkf\" (UniqueName: \"kubernetes.io/projected/7630d135-1664-41e2-9b9e-76a325f47863-kube-api-access-pvhkf\") pod \"glance-default-internal-api-0\" (UID: \"7630d135-1664-41e2-9b9e-76a325f47863\") " pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.294609 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.354794 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.852835 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57d947e-37a6-49b7-bdce-cee499065137" path="/var/lib/kubelet/pods/e57d947e-37a6-49b7-bdce-cee499065137/volumes" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.854137 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7852810-979e-45c7-b5f4-1f6de231699f" path="/var/lib/kubelet/pods/f7852810-979e-45c7-b5f4-1f6de231699f/volumes" Dec 12 08:25:36 crc kubenswrapper[4867]: I1212 08:25:36.920475 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 12 08:25:36 crc kubenswrapper[4867]: W1212 08:25:36.933525 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7630d135_1664_41e2_9b9e_76a325f47863.slice/crio-7c9a95c73e4ff244ae27604e37f40190554c16910782a3706a7adc00b8040672 WatchSource:0}: Error finding container 7c9a95c73e4ff244ae27604e37f40190554c16910782a3706a7adc00b8040672: Status 404 returned error can't find the container with id 7c9a95c73e4ff244ae27604e37f40190554c16910782a3706a7adc00b8040672 Dec 12 08:25:37 crc kubenswrapper[4867]: I1212 08:25:37.081533 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 12 08:25:37 crc kubenswrapper[4867]: W1212 08:25:37.090053 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaea8bd9c_4959_457f_b87e_dff93c91f4e7.slice/crio-858816d83c0f5800bef41389c716021fe16896b0a95b17eff634a1466ff4acca WatchSource:0}: Error finding container 858816d83c0f5800bef41389c716021fe16896b0a95b17eff634a1466ff4acca: Status 404 returned error can't find the container with id 858816d83c0f5800bef41389c716021fe16896b0a95b17eff634a1466ff4acca Dec 12 08:25:37 crc kubenswrapper[4867]: I1212 08:25:37.625185 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aea8bd9c-4959-457f-b87e-dff93c91f4e7","Type":"ContainerStarted","Data":"858816d83c0f5800bef41389c716021fe16896b0a95b17eff634a1466ff4acca"} Dec 12 08:25:37 crc kubenswrapper[4867]: I1212 08:25:37.633795 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7630d135-1664-41e2-9b9e-76a325f47863","Type":"ContainerStarted","Data":"7c9a95c73e4ff244ae27604e37f40190554c16910782a3706a7adc00b8040672"} Dec 12 08:25:38 crc kubenswrapper[4867]: I1212 08:25:38.651115 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aea8bd9c-4959-457f-b87e-dff93c91f4e7","Type":"ContainerStarted","Data":"e1357f9eb537ef3ff717f6b370c32915c8a4b7fb6b60615ea5f49ef075d49134"} Dec 12 08:25:38 crc kubenswrapper[4867]: I1212 08:25:38.652968 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7630d135-1664-41e2-9b9e-76a325f47863","Type":"ContainerStarted","Data":"ab1c7e47be84d4145be90be5a1eaf95c5bebf6e884c4cba5d03759c3c6cc7ba2"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.699264 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerStarted","Data":"c45c539948831f6a2c7314b9f5a259c1250df1bb745656746800276c7aacd025"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.699903 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerStarted","Data":"9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.703534 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerStarted","Data":"0c2ef26e9e0539234a2711588629f1b4c3192e1df0090da32f310134726c47a6"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.703571 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerStarted","Data":"65f20d52bcd69dd154a5e3248e97cb73283ebb3f00a7bb4fdbb759a105e3048d"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.706020 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"aea8bd9c-4959-457f-b87e-dff93c91f4e7","Type":"ContainerStarted","Data":"7f11da7e455a5bf28f8599631e2c6b86604ead7e288217108410d3f008a07761"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.708239 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7630d135-1664-41e2-9b9e-76a325f47863","Type":"ContainerStarted","Data":"092def2e958a3a6a7b68bfa5818061deb1a0744b9770aedaf7b5f682d6b62947"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.710503 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerStarted","Data":"a2fa99ffc154ed9be5c8bcc0ae00db338df02f31962e8ba42e21e1af6ca54af8"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.710529 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerStarted","Data":"fb5b4eaed15ff169d3f777adde32294ac68d72609393a99cbb1208a8b9793899"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.710601 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59b4cf6b67-z2g65" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon-log" containerID="cri-o://fb5b4eaed15ff169d3f777adde32294ac68d72609393a99cbb1208a8b9793899" gracePeriod=30 Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.710714 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-59b4cf6b67-z2g65" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon" containerID="cri-o://a2fa99ffc154ed9be5c8bcc0ae00db338df02f31962e8ba42e21e1af6ca54af8" gracePeriod=30 Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.714903 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerStarted","Data":"e59ee0be4c057d6d38dcb8986bea71a11dfd3dac1f063a428294f0977fd26593"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.714966 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerStarted","Data":"95feedfeb5abc7361bc5625aba08facbe18148dbbf10714d8a55f5cfe6c0c16c"} Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.715111 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64cf5f8fb9-d2m4k" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon-log" containerID="cri-o://95feedfeb5abc7361bc5625aba08facbe18148dbbf10714d8a55f5cfe6c0c16c" gracePeriod=30 Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.715212 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64cf5f8fb9-d2m4k" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon" containerID="cri-o://e59ee0be4c057d6d38dcb8986bea71a11dfd3dac1f063a428294f0977fd26593" gracePeriod=30 Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.729189 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-76dc45d54-sw2qv" podStartSLOduration=2.108786619 podStartE2EDuration="9.729165306s" podCreationTimestamp="2025-12-12 08:25:34 +0000 UTC" firstStartedPulling="2025-12-12 08:25:35.387985121 +0000 UTC m=+5822.959366390" lastFinishedPulling="2025-12-12 08:25:43.008363808 +0000 UTC m=+5830.579745077" observedRunningTime="2025-12-12 08:25:43.719093337 +0000 UTC m=+5831.290474616" watchObservedRunningTime="2025-12-12 08:25:43.729165306 +0000 UTC m=+5831.300546575" Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.772818 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d5d9f84d-gcsp7" podStartSLOduration=2.063755012 podStartE2EDuration="9.772797078s" podCreationTimestamp="2025-12-12 08:25:34 +0000 UTC" firstStartedPulling="2025-12-12 08:25:35.300112922 +0000 UTC m=+5822.871494181" lastFinishedPulling="2025-12-12 08:25:43.009154978 +0000 UTC m=+5830.580536247" observedRunningTime="2025-12-12 08:25:43.747773688 +0000 UTC m=+5831.319154967" watchObservedRunningTime="2025-12-12 08:25:43.772797078 +0000 UTC m=+5831.344178347" Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.773865 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-59b4cf6b67-z2g65" podStartSLOduration=2.272777708 podStartE2EDuration="12.773859735s" podCreationTimestamp="2025-12-12 08:25:31 +0000 UTC" firstStartedPulling="2025-12-12 08:25:32.45279315 +0000 UTC m=+5820.024174419" lastFinishedPulling="2025-12-12 08:25:42.953875187 +0000 UTC m=+5830.525256446" observedRunningTime="2025-12-12 08:25:43.767150648 +0000 UTC m=+5831.338531917" watchObservedRunningTime="2025-12-12 08:25:43.773859735 +0000 UTC m=+5831.345241004" Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.789016 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64cf5f8fb9-d2m4k" podStartSLOduration=2.273360573 podStartE2EDuration="12.78899688s" podCreationTimestamp="2025-12-12 08:25:31 +0000 UTC" firstStartedPulling="2025-12-12 08:25:32.554985075 +0000 UTC m=+5820.126366344" lastFinishedPulling="2025-12-12 08:25:43.070621382 +0000 UTC m=+5830.642002651" observedRunningTime="2025-12-12 08:25:43.785558325 +0000 UTC m=+5831.356939614" watchObservedRunningTime="2025-12-12 08:25:43.78899688 +0000 UTC m=+5831.360378149" Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.804518 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.804500125 podStartE2EDuration="8.804500125s" podCreationTimestamp="2025-12-12 08:25:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:25:43.802866344 +0000 UTC m=+5831.374247613" watchObservedRunningTime="2025-12-12 08:25:43.804500125 +0000 UTC m=+5831.375881394" Dec 12 08:25:43 crc kubenswrapper[4867]: I1212 08:25:43.829081 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.829059614 podStartE2EDuration="8.829059614s" podCreationTimestamp="2025-12-12 08:25:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:25:43.820276686 +0000 UTC m=+5831.391657955" watchObservedRunningTime="2025-12-12 08:25:43.829059614 +0000 UTC m=+5831.400440883" Dec 12 08:25:44 crc kubenswrapper[4867]: I1212 08:25:44.611051 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:44 crc kubenswrapper[4867]: I1212 08:25:44.611539 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:25:44 crc kubenswrapper[4867]: I1212 08:25:44.719429 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:44 crc kubenswrapper[4867]: I1212 08:25:44.719499 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.295282 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.295603 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.325765 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.340687 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.355709 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.355755 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.392409 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.404360 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.743931 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.744946 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.744978 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:46 crc kubenswrapper[4867]: I1212 08:25:46.744993 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:48 crc kubenswrapper[4867]: I1212 08:25:48.762694 4867 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 12 08:25:48 crc kubenswrapper[4867]: I1212 08:25:48.767288 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 08:25:48 crc kubenswrapper[4867]: I1212 08:25:48.861465 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:50 crc kubenswrapper[4867]: I1212 08:25:50.789958 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 12 08:25:50 crc kubenswrapper[4867]: I1212 08:25:50.880373 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 12 08:25:51 crc kubenswrapper[4867]: I1212 08:25:51.758212 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:25:51 crc kubenswrapper[4867]: I1212 08:25:51.858977 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:25:54 crc kubenswrapper[4867]: I1212 08:25:54.598434 4867 scope.go:117] "RemoveContainer" containerID="73cd0d4fc21d8b40133a343b64467b830c47f69e8b79463d7de44050ee16cf0c" Dec 12 08:25:54 crc kubenswrapper[4867]: I1212 08:25:54.613352 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.98:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.98:8443: connect: connection refused" Dec 12 08:25:54 crc kubenswrapper[4867]: I1212 08:25:54.641871 4867 scope.go:117] "RemoveContainer" containerID="2868dd45bea43f598497baea5c30c24ca76d1a19d4a00a7dd491dc9990cedf06" Dec 12 08:25:54 crc kubenswrapper[4867]: I1212 08:25:54.721933 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.99:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.99:8443: connect: connection refused" Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.046962 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8af0-account-create-update-nmdss"] Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.060004 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-2mpzh"] Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.069722 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8af0-account-create-update-nmdss"] Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.079200 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-2mpzh"] Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.850545 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5764dd52-d7f4-483f-93c9-fd5a70ef6a1d" path="/var/lib/kubelet/pods/5764dd52-d7f4-483f-93c9-fd5a70ef6a1d/volumes" Dec 12 08:26:02 crc kubenswrapper[4867]: I1212 08:26:02.851795 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9212e1c2-e6d1-463a-b33b-323d72d37fec" path="/var/lib/kubelet/pods/9212e1c2-e6d1-463a-b33b-323d72d37fec/volumes" Dec 12 08:26:06 crc kubenswrapper[4867]: I1212 08:26:06.539717 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:26:06 crc kubenswrapper[4867]: I1212 08:26:06.569353 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.381145 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.406905 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.458396 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.794160 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.796645 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.806955 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.950912 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.951645 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.951971 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4k2\" (UniqueName: \"kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.974989 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" containerID="cri-o://c45c539948831f6a2c7314b9f5a259c1250df1bb745656746800276c7aacd025" gracePeriod=30 Dec 12 08:26:08 crc kubenswrapper[4867]: I1212 08:26:08.975012 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon-log" containerID="cri-o://9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6" gracePeriod=30 Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.054049 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.054489 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4k2\" (UniqueName: \"kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.054602 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.054657 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.055192 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.079146 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4k2\" (UniqueName: \"kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2\") pod \"community-operators-kx6rs\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.117315 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.932895 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:09 crc kubenswrapper[4867]: I1212 08:26:09.985937 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerStarted","Data":"9bb9d3a0d958021d2e805482c00010d25dc796d56d3048ceb09f632662eef581"} Dec 12 08:26:10 crc kubenswrapper[4867]: I1212 08:26:10.996220 4867 generic.go:334] "Generic (PLEG): container finished" podID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerID="6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94" exitCode=0 Dec 12 08:26:10 crc kubenswrapper[4867]: I1212 08:26:10.996398 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerDied","Data":"6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94"} Dec 12 08:26:12 crc kubenswrapper[4867]: I1212 08:26:12.008448 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerStarted","Data":"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc"} Dec 12 08:26:13 crc kubenswrapper[4867]: I1212 08:26:13.017296 4867 generic.go:334] "Generic (PLEG): container finished" podID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerID="bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc" exitCode=0 Dec 12 08:26:13 crc kubenswrapper[4867]: I1212 08:26:13.017337 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerDied","Data":"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc"} Dec 12 08:26:13 crc kubenswrapper[4867]: I1212 08:26:13.019455 4867 generic.go:334] "Generic (PLEG): container finished" podID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerID="c45c539948831f6a2c7314b9f5a259c1250df1bb745656746800276c7aacd025" exitCode=0 Dec 12 08:26:13 crc kubenswrapper[4867]: I1212 08:26:13.019481 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerDied","Data":"c45c539948831f6a2c7314b9f5a259c1250df1bb745656746800276c7aacd025"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.033910 4867 generic.go:334] "Generic (PLEG): container finished" podID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerID="e59ee0be4c057d6d38dcb8986bea71a11dfd3dac1f063a428294f0977fd26593" exitCode=137 Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.034272 4867 generic.go:334] "Generic (PLEG): container finished" podID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerID="95feedfeb5abc7361bc5625aba08facbe18148dbbf10714d8a55f5cfe6c0c16c" exitCode=137 Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.034013 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerDied","Data":"e59ee0be4c057d6d38dcb8986bea71a11dfd3dac1f063a428294f0977fd26593"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.034364 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerDied","Data":"95feedfeb5abc7361bc5625aba08facbe18148dbbf10714d8a55f5cfe6c0c16c"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.044111 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerStarted","Data":"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.047623 4867 generic.go:334] "Generic (PLEG): container finished" podID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerID="a2fa99ffc154ed9be5c8bcc0ae00db338df02f31962e8ba42e21e1af6ca54af8" exitCode=137 Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.047658 4867 generic.go:334] "Generic (PLEG): container finished" podID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerID="fb5b4eaed15ff169d3f777adde32294ac68d72609393a99cbb1208a8b9793899" exitCode=137 Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.047684 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerDied","Data":"a2fa99ffc154ed9be5c8bcc0ae00db338df02f31962e8ba42e21e1af6ca54af8"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.047713 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerDied","Data":"fb5b4eaed15ff169d3f777adde32294ac68d72609393a99cbb1208a8b9793899"} Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.068019 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kx6rs" podStartSLOduration=3.544032024 podStartE2EDuration="6.067996346s" podCreationTimestamp="2025-12-12 08:26:08 +0000 UTC" firstStartedPulling="2025-12-12 08:26:10.997915659 +0000 UTC m=+5858.569296938" lastFinishedPulling="2025-12-12 08:26:13.521879991 +0000 UTC m=+5861.093261260" observedRunningTime="2025-12-12 08:26:14.064707475 +0000 UTC m=+5861.636102974" watchObservedRunningTime="2025-12-12 08:26:14.067996346 +0000 UTC m=+5861.639377615" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.148499 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.219589 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288047 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr2xp\" (UniqueName: \"kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp\") pod \"491be33e-0b7f-4408-97bf-f235a4c6f824\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288151 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts\") pod \"491be33e-0b7f-4408-97bf-f235a4c6f824\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288300 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs\") pod \"491be33e-0b7f-4408-97bf-f235a4c6f824\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288326 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data\") pod \"491be33e-0b7f-4408-97bf-f235a4c6f824\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288394 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key\") pod \"491be33e-0b7f-4408-97bf-f235a4c6f824\" (UID: \"491be33e-0b7f-4408-97bf-f235a4c6f824\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.288974 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs" (OuterVolumeSpecName: "logs") pod "491be33e-0b7f-4408-97bf-f235a4c6f824" (UID: "491be33e-0b7f-4408-97bf-f235a4c6f824"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.298555 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "491be33e-0b7f-4408-97bf-f235a4c6f824" (UID: "491be33e-0b7f-4408-97bf-f235a4c6f824"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.298574 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp" (OuterVolumeSpecName: "kube-api-access-hr2xp") pod "491be33e-0b7f-4408-97bf-f235a4c6f824" (UID: "491be33e-0b7f-4408-97bf-f235a4c6f824"). InnerVolumeSpecName "kube-api-access-hr2xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.323244 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts" (OuterVolumeSpecName: "scripts") pod "491be33e-0b7f-4408-97bf-f235a4c6f824" (UID: "491be33e-0b7f-4408-97bf-f235a4c6f824"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.324415 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data" (OuterVolumeSpecName: "config-data") pod "491be33e-0b7f-4408-97bf-f235a4c6f824" (UID: "491be33e-0b7f-4408-97bf-f235a4c6f824"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.390705 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key\") pod \"9493bb26-5213-4a7a-afb9-fa172e800f0c\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.391048 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data\") pod \"9493bb26-5213-4a7a-afb9-fa172e800f0c\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.391216 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs\") pod \"9493bb26-5213-4a7a-afb9-fa172e800f0c\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.391356 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts\") pod \"9493bb26-5213-4a7a-afb9-fa172e800f0c\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.391480 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-564mx\" (UniqueName: \"kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx\") pod \"9493bb26-5213-4a7a-afb9-fa172e800f0c\" (UID: \"9493bb26-5213-4a7a-afb9-fa172e800f0c\") " Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.391824 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs" (OuterVolumeSpecName: "logs") pod "9493bb26-5213-4a7a-afb9-fa172e800f0c" (UID: "9493bb26-5213-4a7a-afb9-fa172e800f0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392347 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr2xp\" (UniqueName: \"kubernetes.io/projected/491be33e-0b7f-4408-97bf-f235a4c6f824-kube-api-access-hr2xp\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392456 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392535 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9493bb26-5213-4a7a-afb9-fa172e800f0c-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392627 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/491be33e-0b7f-4408-97bf-f235a4c6f824-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392717 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/491be33e-0b7f-4408-97bf-f235a4c6f824-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.392916 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/491be33e-0b7f-4408-97bf-f235a4c6f824-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.395337 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9493bb26-5213-4a7a-afb9-fa172e800f0c" (UID: "9493bb26-5213-4a7a-afb9-fa172e800f0c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.395826 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx" (OuterVolumeSpecName: "kube-api-access-564mx") pod "9493bb26-5213-4a7a-afb9-fa172e800f0c" (UID: "9493bb26-5213-4a7a-afb9-fa172e800f0c"). InnerVolumeSpecName "kube-api-access-564mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.414899 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts" (OuterVolumeSpecName: "scripts") pod "9493bb26-5213-4a7a-afb9-fa172e800f0c" (UID: "9493bb26-5213-4a7a-afb9-fa172e800f0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.418994 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data" (OuterVolumeSpecName: "config-data") pod "9493bb26-5213-4a7a-afb9-fa172e800f0c" (UID: "9493bb26-5213-4a7a-afb9-fa172e800f0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.494703 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9493bb26-5213-4a7a-afb9-fa172e800f0c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.494760 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.494772 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9493bb26-5213-4a7a-afb9-fa172e800f0c-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.494788 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-564mx\" (UniqueName: \"kubernetes.io/projected/9493bb26-5213-4a7a-afb9-fa172e800f0c-kube-api-access-564mx\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:14 crc kubenswrapper[4867]: I1212 08:26:14.611780 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.98:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.98:8443: connect: connection refused" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.059290 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-59b4cf6b67-z2g65" event={"ID":"491be33e-0b7f-4408-97bf-f235a4c6f824","Type":"ContainerDied","Data":"dda0b3c36b0c2a30018389074827c1019b09c3a7e38b76e18b99718bbdda553a"} Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.059330 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-59b4cf6b67-z2g65" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.059349 4867 scope.go:117] "RemoveContainer" containerID="a2fa99ffc154ed9be5c8bcc0ae00db338df02f31962e8ba42e21e1af6ca54af8" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.064987 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64cf5f8fb9-d2m4k" event={"ID":"9493bb26-5213-4a7a-afb9-fa172e800f0c","Type":"ContainerDied","Data":"bdec6494b6f769600e3cb8a0ac6750c3a4474cc743e72ec0983a22e481bd1b1a"} Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.065001 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64cf5f8fb9-d2m4k" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.083749 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.097290 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-59b4cf6b67-z2g65"] Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.132134 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.143758 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64cf5f8fb9-d2m4k"] Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.251831 4867 scope.go:117] "RemoveContainer" containerID="fb5b4eaed15ff169d3f777adde32294ac68d72609393a99cbb1208a8b9793899" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.282904 4867 scope.go:117] "RemoveContainer" containerID="e59ee0be4c057d6d38dcb8986bea71a11dfd3dac1f063a428294f0977fd26593" Dec 12 08:26:15 crc kubenswrapper[4867]: I1212 08:26:15.469458 4867 scope.go:117] "RemoveContainer" containerID="95feedfeb5abc7361bc5625aba08facbe18148dbbf10714d8a55f5cfe6c0c16c" Dec 12 08:26:16 crc kubenswrapper[4867]: I1212 08:26:16.034388 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2rzp2"] Dec 12 08:26:16 crc kubenswrapper[4867]: I1212 08:26:16.043831 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2rzp2"] Dec 12 08:26:16 crc kubenswrapper[4867]: I1212 08:26:16.851800 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" path="/var/lib/kubelet/pods/491be33e-0b7f-4408-97bf-f235a4c6f824/volumes" Dec 12 08:26:16 crc kubenswrapper[4867]: I1212 08:26:16.853296 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71445001-629a-4b2c-a78f-428aa0164958" path="/var/lib/kubelet/pods/71445001-629a-4b2c-a78f-428aa0164958/volumes" Dec 12 08:26:16 crc kubenswrapper[4867]: I1212 08:26:16.854108 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" path="/var/lib/kubelet/pods/9493bb26-5213-4a7a-afb9-fa172e800f0c/volumes" Dec 12 08:26:19 crc kubenswrapper[4867]: I1212 08:26:19.121333 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:19 crc kubenswrapper[4867]: I1212 08:26:19.121496 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:19 crc kubenswrapper[4867]: I1212 08:26:19.170042 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:20 crc kubenswrapper[4867]: I1212 08:26:20.164339 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:20 crc kubenswrapper[4867]: I1212 08:26:20.221083 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.132050 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kx6rs" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="registry-server" containerID="cri-o://3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3" gracePeriod=2 Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.595370 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.774852 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content\") pod \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.774952 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities\") pod \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.774992 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw4k2\" (UniqueName: \"kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2\") pod \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\" (UID: \"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a\") " Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.776785 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities" (OuterVolumeSpecName: "utilities") pod "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" (UID: "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.780692 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2" (OuterVolumeSpecName: "kube-api-access-pw4k2") pod "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" (UID: "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a"). InnerVolumeSpecName "kube-api-access-pw4k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.831694 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" (UID: "e9e89797-4e63-4b6b-98a4-7bc031aa1b7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.877473 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.877514 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:22 crc kubenswrapper[4867]: I1212 08:26:22.877528 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw4k2\" (UniqueName: \"kubernetes.io/projected/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a-kube-api-access-pw4k2\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.143408 4867 generic.go:334] "Generic (PLEG): container finished" podID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerID="3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3" exitCode=0 Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.143456 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerDied","Data":"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3"} Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.143483 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kx6rs" event={"ID":"e9e89797-4e63-4b6b-98a4-7bc031aa1b7a","Type":"ContainerDied","Data":"9bb9d3a0d958021d2e805482c00010d25dc796d56d3048ceb09f632662eef581"} Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.143502 4867 scope.go:117] "RemoveContainer" containerID="3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.143510 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kx6rs" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.170268 4867 scope.go:117] "RemoveContainer" containerID="bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.175320 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.185629 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kx6rs"] Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.197679 4867 scope.go:117] "RemoveContainer" containerID="6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.244872 4867 scope.go:117] "RemoveContainer" containerID="3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3" Dec 12 08:26:23 crc kubenswrapper[4867]: E1212 08:26:23.245434 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3\": container with ID starting with 3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3 not found: ID does not exist" containerID="3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.245489 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3"} err="failed to get container status \"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3\": rpc error: code = NotFound desc = could not find container \"3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3\": container with ID starting with 3f7f80c7c2dce7f11bcf6b90d9803087f1859e3af4e9a536acdc550a5b01fbd3 not found: ID does not exist" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.245526 4867 scope.go:117] "RemoveContainer" containerID="bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc" Dec 12 08:26:23 crc kubenswrapper[4867]: E1212 08:26:23.245992 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc\": container with ID starting with bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc not found: ID does not exist" containerID="bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.246033 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc"} err="failed to get container status \"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc\": rpc error: code = NotFound desc = could not find container \"bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc\": container with ID starting with bdf6562e7c57e37563d0424bb40902b681694f425632f36474e038ece46488cc not found: ID does not exist" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.246053 4867 scope.go:117] "RemoveContainer" containerID="6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94" Dec 12 08:26:23 crc kubenswrapper[4867]: E1212 08:26:23.246767 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94\": container with ID starting with 6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94 not found: ID does not exist" containerID="6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94" Dec 12 08:26:23 crc kubenswrapper[4867]: I1212 08:26:23.246820 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94"} err="failed to get container status \"6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94\": rpc error: code = NotFound desc = could not find container \"6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94\": container with ID starting with 6c96aa113a4561e50be393bd31088976b44b0a1cae69c78ec0ef29bb1eed6d94 not found: ID does not exist" Dec 12 08:26:24 crc kubenswrapper[4867]: I1212 08:26:24.611557 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.98:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.98:8443: connect: connection refused" Dec 12 08:26:24 crc kubenswrapper[4867]: I1212 08:26:24.849098 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" path="/var/lib/kubelet/pods/e9e89797-4e63-4b6b-98a4-7bc031aa1b7a/volumes" Dec 12 08:26:34 crc kubenswrapper[4867]: I1212 08:26:34.612349 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-76dc45d54-sw2qv" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.98:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.98:8443: connect: connection refused" Dec 12 08:26:34 crc kubenswrapper[4867]: I1212 08:26:34.613501 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:26:39 crc kubenswrapper[4867]: E1212 08:26:39.240783 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1e86b4f_3c98_4a37_8e2f_866551494bad.slice/crio-conmon-9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.290497 4867 generic.go:334] "Generic (PLEG): container finished" podID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerID="9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6" exitCode=137 Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.290544 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerDied","Data":"9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6"} Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.290567 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-76dc45d54-sw2qv" event={"ID":"b1e86b4f-3c98-4a37-8e2f-866551494bad","Type":"ContainerDied","Data":"b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c"} Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.290579 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9b0d6d9e9a938aec7a7d8ff50967b39eb8adf5ecdb612fee92bca0eaa02349c" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.308638 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416202 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416303 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416382 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsl55\" (UniqueName: \"kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416456 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416558 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416596 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.416626 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key\") pod \"b1e86b4f-3c98-4a37-8e2f-866551494bad\" (UID: \"b1e86b4f-3c98-4a37-8e2f-866551494bad\") " Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.417335 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs" (OuterVolumeSpecName: "logs") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.418417 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1e86b4f-3c98-4a37-8e2f-866551494bad-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.421609 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.429471 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55" (OuterVolumeSpecName: "kube-api-access-hsl55") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "kube-api-access-hsl55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.442138 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts" (OuterVolumeSpecName: "scripts") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.443306 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.447428 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data" (OuterVolumeSpecName: "config-data") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.471007 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "b1e86b4f-3c98-4a37-8e2f-866551494bad" (UID: "b1e86b4f-3c98-4a37-8e2f-866551494bad"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519628 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsl55\" (UniqueName: \"kubernetes.io/projected/b1e86b4f-3c98-4a37-8e2f-866551494bad-kube-api-access-hsl55\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519665 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519682 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519695 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519705 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e86b4f-3c98-4a37-8e2f-866551494bad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:39 crc kubenswrapper[4867]: I1212 08:26:39.519714 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1e86b4f-3c98-4a37-8e2f-866551494bad-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:40 crc kubenswrapper[4867]: I1212 08:26:40.298676 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-76dc45d54-sw2qv" Dec 12 08:26:40 crc kubenswrapper[4867]: I1212 08:26:40.336833 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:26:40 crc kubenswrapper[4867]: I1212 08:26:40.345132 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-76dc45d54-sw2qv"] Dec 12 08:26:40 crc kubenswrapper[4867]: I1212 08:26:40.849278 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" path="/var/lib/kubelet/pods/b1e86b4f-3c98-4a37-8e2f-866551494bad/volumes" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.641567 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f8df8656d-5rfkc"] Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642770 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642790 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642818 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642825 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642837 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642847 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642862 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642870 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642883 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="registry-server" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642890 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="registry-server" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642903 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642912 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642938 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="extract-utilities" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642946 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="extract-utilities" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642954 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="extract-content" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642960 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="extract-content" Dec 12 08:26:48 crc kubenswrapper[4867]: E1212 08:26:48.642984 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.642992 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643246 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643265 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9e89797-4e63-4b6b-98a4-7bc031aa1b7a" containerName="registry-server" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643281 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643290 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="491be33e-0b7f-4408-97bf-f235a4c6f824" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643305 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9493bb26-5213-4a7a-afb9-fa172e800f0c" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643317 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.643334 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e86b4f-3c98-4a37-8e2f-866551494bad" containerName="horizon-log" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.644352 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.651785 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8df8656d-5rfkc"] Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801550 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phvk5\" (UniqueName: \"kubernetes.io/projected/826ddec1-f8e5-41fc-bf66-a58f07371c55-kube-api-access-phvk5\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801631 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/826ddec1-f8e5-41fc-bf66-a58f07371c55-logs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801692 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-secret-key\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801715 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-scripts\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801735 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-config-data\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801760 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-tls-certs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.801835 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-combined-ca-bundle\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903669 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phvk5\" (UniqueName: \"kubernetes.io/projected/826ddec1-f8e5-41fc-bf66-a58f07371c55-kube-api-access-phvk5\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903743 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/826ddec1-f8e5-41fc-bf66-a58f07371c55-logs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903787 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-secret-key\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903805 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-scripts\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903819 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-config-data\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903839 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-tls-certs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.903874 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-combined-ca-bundle\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.904310 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/826ddec1-f8e5-41fc-bf66-a58f07371c55-logs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.905046 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-scripts\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.905631 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/826ddec1-f8e5-41fc-bf66-a58f07371c55-config-data\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.909500 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-secret-key\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.909511 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-combined-ca-bundle\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.916445 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/826ddec1-f8e5-41fc-bf66-a58f07371c55-horizon-tls-certs\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.920965 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phvk5\" (UniqueName: \"kubernetes.io/projected/826ddec1-f8e5-41fc-bf66-a58f07371c55-kube-api-access-phvk5\") pod \"horizon-f8df8656d-5rfkc\" (UID: \"826ddec1-f8e5-41fc-bf66-a58f07371c55\") " pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:48 crc kubenswrapper[4867]: I1212 08:26:48.968289 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:49 crc kubenswrapper[4867]: I1212 08:26:49.058385 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pjv5k"] Dec 12 08:26:49 crc kubenswrapper[4867]: I1212 08:26:49.070395 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-fe9d-account-create-update-hc4nz"] Dec 12 08:26:49 crc kubenswrapper[4867]: I1212 08:26:49.079646 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pjv5k"] Dec 12 08:26:49 crc kubenswrapper[4867]: I1212 08:26:49.098005 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-fe9d-account-create-update-hc4nz"] Dec 12 08:26:49 crc kubenswrapper[4867]: I1212 08:26:49.476151 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8df8656d-5rfkc"] Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.012143 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pnzm6"] Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.013591 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.027902 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pnzm6"] Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.039785 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-2247-account-create-update-gtr66"] Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.041200 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.044308 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.061834 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2247-account-create-update-gtr66"] Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.125330 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55426\" (UniqueName: \"kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.125565 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.125780 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqjs\" (UniqueName: \"kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.125829 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.227202 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.227397 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqjs\" (UniqueName: \"kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.227430 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.227505 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55426\" (UniqueName: \"kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.227935 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.228436 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.245347 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqjs\" (UniqueName: \"kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs\") pod \"heat-2247-account-create-update-gtr66\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.245603 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55426\" (UniqueName: \"kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426\") pod \"heat-db-create-pnzm6\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.359392 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.373131 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.391791 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8df8656d-5rfkc" event={"ID":"826ddec1-f8e5-41fc-bf66-a58f07371c55","Type":"ContainerStarted","Data":"7a2e3274f28ce050e06f74fcdc9e59adc6588731a32a6825bb096b1a59a9319b"} Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.391854 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8df8656d-5rfkc" event={"ID":"826ddec1-f8e5-41fc-bf66-a58f07371c55","Type":"ContainerStarted","Data":"89006d3a52f6a5ab2d5af86477ff62f5837ec2fadc0e8cf4b4ff759bc1415481"} Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.391864 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8df8656d-5rfkc" event={"ID":"826ddec1-f8e5-41fc-bf66-a58f07371c55","Type":"ContainerStarted","Data":"9ea8ddf094c13b65c9d4c6657e9a731b252167deadd7998e5b79b7944c9f521b"} Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.424687 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f8df8656d-5rfkc" podStartSLOduration=2.424642986 podStartE2EDuration="2.424642986s" podCreationTimestamp="2025-12-12 08:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:26:50.410382291 +0000 UTC m=+5897.981763570" watchObservedRunningTime="2025-12-12 08:26:50.424642986 +0000 UTC m=+5897.996024255" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.853623 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3475cd4e-37de-44b2-8f2a-fb4f370b5a89" path="/var/lib/kubelet/pods/3475cd4e-37de-44b2-8f2a-fb4f370b5a89/volumes" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.854794 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03" path="/var/lib/kubelet/pods/b1c3f281-8e81-45ef-9ab6-cd7b1a9c4c03/volumes" Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.874935 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pnzm6"] Dec 12 08:26:50 crc kubenswrapper[4867]: W1212 08:26:50.883657 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d33ce8e_3c6a_40ef_b7e3_11e36b73eb3e.slice/crio-e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823 WatchSource:0}: Error finding container e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823: Status 404 returned error can't find the container with id e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823 Dec 12 08:26:50 crc kubenswrapper[4867]: I1212 08:26:50.889910 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2247-account-create-update-gtr66"] Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.402769 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pnzm6" event={"ID":"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e","Type":"ContainerStarted","Data":"02430ea5aa82557393d2a4654b9ce856be20ad2afc41385b3e21992d93e523c5"} Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.402811 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pnzm6" event={"ID":"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e","Type":"ContainerStarted","Data":"e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823"} Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.406002 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2247-account-create-update-gtr66" event={"ID":"abaeb799-309d-4a35-9e9b-cdcfe0a4065f","Type":"ContainerStarted","Data":"e4e4d824a2401411622e7fb861727ad1415ec2046eaaf106644e9e3ec9a9f778"} Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.406063 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2247-account-create-update-gtr66" event={"ID":"abaeb799-309d-4a35-9e9b-cdcfe0a4065f","Type":"ContainerStarted","Data":"5a71a68cafb7d8326b313f38c5f6f90e3817c3b4d1610088e6082068ca34ded5"} Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.421428 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-pnzm6" podStartSLOduration=2.421411848 podStartE2EDuration="2.421411848s" podCreationTimestamp="2025-12-12 08:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:26:51.419843129 +0000 UTC m=+5898.991224398" watchObservedRunningTime="2025-12-12 08:26:51.421411848 +0000 UTC m=+5898.992793117" Dec 12 08:26:51 crc kubenswrapper[4867]: I1212 08:26:51.442171 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-2247-account-create-update-gtr66" podStartSLOduration=2.442148512 podStartE2EDuration="2.442148512s" podCreationTimestamp="2025-12-12 08:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:26:51.436357199 +0000 UTC m=+5899.007738478" watchObservedRunningTime="2025-12-12 08:26:51.442148512 +0000 UTC m=+5899.013529791" Dec 12 08:26:52 crc kubenswrapper[4867]: I1212 08:26:52.429451 4867 generic.go:334] "Generic (PLEG): container finished" podID="7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" containerID="02430ea5aa82557393d2a4654b9ce856be20ad2afc41385b3e21992d93e523c5" exitCode=0 Dec 12 08:26:52 crc kubenswrapper[4867]: I1212 08:26:52.429620 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pnzm6" event={"ID":"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e","Type":"ContainerDied","Data":"02430ea5aa82557393d2a4654b9ce856be20ad2afc41385b3e21992d93e523c5"} Dec 12 08:26:52 crc kubenswrapper[4867]: I1212 08:26:52.434254 4867 generic.go:334] "Generic (PLEG): container finished" podID="abaeb799-309d-4a35-9e9b-cdcfe0a4065f" containerID="e4e4d824a2401411622e7fb861727ad1415ec2046eaaf106644e9e3ec9a9f778" exitCode=0 Dec 12 08:26:52 crc kubenswrapper[4867]: I1212 08:26:52.434307 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2247-account-create-update-gtr66" event={"ID":"abaeb799-309d-4a35-9e9b-cdcfe0a4065f","Type":"ContainerDied","Data":"e4e4d824a2401411622e7fb861727ad1415ec2046eaaf106644e9e3ec9a9f778"} Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.837794 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.845291 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.900028 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsqjs\" (UniqueName: \"kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs\") pod \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.900311 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55426\" (UniqueName: \"kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426\") pod \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.900457 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts\") pod \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\" (UID: \"abaeb799-309d-4a35-9e9b-cdcfe0a4065f\") " Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.900492 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts\") pod \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\" (UID: \"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e\") " Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.901690 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" (UID: "7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.901753 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "abaeb799-309d-4a35-9e9b-cdcfe0a4065f" (UID: "abaeb799-309d-4a35-9e9b-cdcfe0a4065f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.906412 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426" (OuterVolumeSpecName: "kube-api-access-55426") pod "7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" (UID: "7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e"). InnerVolumeSpecName "kube-api-access-55426". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:53 crc kubenswrapper[4867]: I1212 08:26:53.906574 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs" (OuterVolumeSpecName: "kube-api-access-hsqjs") pod "abaeb799-309d-4a35-9e9b-cdcfe0a4065f" (UID: "abaeb799-309d-4a35-9e9b-cdcfe0a4065f"). InnerVolumeSpecName "kube-api-access-hsqjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.002530 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55426\" (UniqueName: \"kubernetes.io/projected/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-kube-api-access-55426\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.002580 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.002592 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.002608 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsqjs\" (UniqueName: \"kubernetes.io/projected/abaeb799-309d-4a35-9e9b-cdcfe0a4065f-kube-api-access-hsqjs\") on node \"crc\" DevicePath \"\"" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.453428 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pnzm6" event={"ID":"7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e","Type":"ContainerDied","Data":"e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823"} Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.453847 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7013dbaef2892e5e89767ce3ad02b5003931682fd2cd173a6adf70067116823" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.453443 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pnzm6" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.455618 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2247-account-create-update-gtr66" event={"ID":"abaeb799-309d-4a35-9e9b-cdcfe0a4065f","Type":"ContainerDied","Data":"5a71a68cafb7d8326b313f38c5f6f90e3817c3b4d1610088e6082068ca34ded5"} Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.455649 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a71a68cafb7d8326b313f38c5f6f90e3817c3b4d1610088e6082068ca34ded5" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.455702 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2247-account-create-update-gtr66" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.754629 4867 scope.go:117] "RemoveContainer" containerID="0d90299f10387afb4d7c2724416a7468daf60a136d919a64c99eb93ffc0fa997" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.783808 4867 scope.go:117] "RemoveContainer" containerID="1b46710b606b3e148a086b597a61d7e794f4ce1500c5ba0aa27b018ef8624166" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.833753 4867 scope.go:117] "RemoveContainer" containerID="0a8e5392a6aa0f51d7bce68e67536728957ae4533f9397c7491203c9c7bfd8d4" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.886166 4867 scope.go:117] "RemoveContainer" containerID="5dfad9aee108d994e453cf359b424a8cb7c3d9aa73fb060ecb090f17b311b9c1" Dec 12 08:26:54 crc kubenswrapper[4867]: I1212 08:26:54.933917 4867 scope.go:117] "RemoveContainer" containerID="d0f25ff5dc033f4df4696b0cde0968038e55cb17617a44e5a10d86985545c6fc" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.067562 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-xgdqt"] Dec 12 08:26:55 crc kubenswrapper[4867]: E1212 08:26:55.067943 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" containerName="mariadb-database-create" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.067961 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" containerName="mariadb-database-create" Dec 12 08:26:55 crc kubenswrapper[4867]: E1212 08:26:55.067981 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaeb799-309d-4a35-9e9b-cdcfe0a4065f" containerName="mariadb-account-create-update" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.067989 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaeb799-309d-4a35-9e9b-cdcfe0a4065f" containerName="mariadb-account-create-update" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.068193 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" containerName="mariadb-database-create" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.068213 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaeb799-309d-4a35-9e9b-cdcfe0a4065f" containerName="mariadb-account-create-update" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.070709 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.072465 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qcm4r" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.073243 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.079625 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xgdqt"] Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.125531 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.127190 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.127295 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqr7l\" (UniqueName: \"kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.230115 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.233184 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqr7l\" (UniqueName: \"kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.233509 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.235920 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.243652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.253736 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqr7l\" (UniqueName: \"kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l\") pod \"heat-db-sync-xgdqt\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.391597 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xgdqt" Dec 12 08:26:55 crc kubenswrapper[4867]: I1212 08:26:55.875883 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-xgdqt"] Dec 12 08:26:56 crc kubenswrapper[4867]: I1212 08:26:56.474153 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xgdqt" event={"ID":"0fae1753-5101-4148-a0ca-20a48294a726","Type":"ContainerStarted","Data":"05d099b4f16b436e439be86aeb2c3271e99291e80bc7d9d26b815430c5b482ff"} Dec 12 08:26:57 crc kubenswrapper[4867]: I1212 08:26:57.041987 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-j67rb"] Dec 12 08:26:57 crc kubenswrapper[4867]: I1212 08:26:57.053184 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-j67rb"] Dec 12 08:26:58 crc kubenswrapper[4867]: I1212 08:26:58.849610 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82193c94-00e1-47d0-ab01-992d27297d32" path="/var/lib/kubelet/pods/82193c94-00e1-47d0-ab01-992d27297d32/volumes" Dec 12 08:26:58 crc kubenswrapper[4867]: I1212 08:26:58.968482 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:26:58 crc kubenswrapper[4867]: I1212 08:26:58.970110 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:27:05 crc kubenswrapper[4867]: I1212 08:27:05.557806 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xgdqt" event={"ID":"0fae1753-5101-4148-a0ca-20a48294a726","Type":"ContainerStarted","Data":"ef41feb6a82e9c2d1d78602fa8f55847064fe9a7d708b0d7fed7d52217855a01"} Dec 12 08:27:05 crc kubenswrapper[4867]: I1212 08:27:05.580599 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-xgdqt" podStartSLOduration=2.097956253 podStartE2EDuration="10.580576446s" podCreationTimestamp="2025-12-12 08:26:55 +0000 UTC" firstStartedPulling="2025-12-12 08:26:55.896638417 +0000 UTC m=+5903.468019686" lastFinishedPulling="2025-12-12 08:27:04.37925861 +0000 UTC m=+5911.950639879" observedRunningTime="2025-12-12 08:27:05.57589765 +0000 UTC m=+5913.147278949" watchObservedRunningTime="2025-12-12 08:27:05.580576446 +0000 UTC m=+5913.151957715" Dec 12 08:27:06 crc kubenswrapper[4867]: I1212 08:27:06.570269 4867 generic.go:334] "Generic (PLEG): container finished" podID="0fae1753-5101-4148-a0ca-20a48294a726" containerID="ef41feb6a82e9c2d1d78602fa8f55847064fe9a7d708b0d7fed7d52217855a01" exitCode=0 Dec 12 08:27:06 crc kubenswrapper[4867]: I1212 08:27:06.570445 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xgdqt" event={"ID":"0fae1753-5101-4148-a0ca-20a48294a726","Type":"ContainerDied","Data":"ef41feb6a82e9c2d1d78602fa8f55847064fe9a7d708b0d7fed7d52217855a01"} Dec 12 08:27:07 crc kubenswrapper[4867]: I1212 08:27:07.943420 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xgdqt" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.063776 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle\") pod \"0fae1753-5101-4148-a0ca-20a48294a726\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.064143 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data\") pod \"0fae1753-5101-4148-a0ca-20a48294a726\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.064258 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqr7l\" (UniqueName: \"kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l\") pod \"0fae1753-5101-4148-a0ca-20a48294a726\" (UID: \"0fae1753-5101-4148-a0ca-20a48294a726\") " Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.070016 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l" (OuterVolumeSpecName: "kube-api-access-bqr7l") pod "0fae1753-5101-4148-a0ca-20a48294a726" (UID: "0fae1753-5101-4148-a0ca-20a48294a726"). InnerVolumeSpecName "kube-api-access-bqr7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.097037 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fae1753-5101-4148-a0ca-20a48294a726" (UID: "0fae1753-5101-4148-a0ca-20a48294a726"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.134372 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data" (OuterVolumeSpecName: "config-data") pod "0fae1753-5101-4148-a0ca-20a48294a726" (UID: "0fae1753-5101-4148-a0ca-20a48294a726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.166246 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.166282 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqr7l\" (UniqueName: \"kubernetes.io/projected/0fae1753-5101-4148-a0ca-20a48294a726-kube-api-access-bqr7l\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.166293 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fae1753-5101-4148-a0ca-20a48294a726-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.592608 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-xgdqt" event={"ID":"0fae1753-5101-4148-a0ca-20a48294a726","Type":"ContainerDied","Data":"05d099b4f16b436e439be86aeb2c3271e99291e80bc7d9d26b815430c5b482ff"} Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.592856 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05d099b4f16b436e439be86aeb2c3271e99291e80bc7d9d26b815430c5b482ff" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.592916 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-xgdqt" Dec 12 08:27:08 crc kubenswrapper[4867]: I1212 08:27:08.969684 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f8df8656d-5rfkc" podUID="826ddec1-f8e5-41fc-bf66-a58f07371c55" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.103:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.103:8443: connect: connection refused" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.781827 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:09 crc kubenswrapper[4867]: E1212 08:27:09.782611 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fae1753-5101-4148-a0ca-20a48294a726" containerName="heat-db-sync" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.782628 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fae1753-5101-4148-a0ca-20a48294a726" containerName="heat-db-sync" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.782864 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fae1753-5101-4148-a0ca-20a48294a726" containerName="heat-db-sync" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.783576 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.788648 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.791848 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-qcm4r" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.801156 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.827075 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.851742 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.853399 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.859792 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.892941 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.904329 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.904366 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.904464 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.904586 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs6jf\" (UniqueName: \"kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.912820 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.917350 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.922919 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 12 08:27:09 crc kubenswrapper[4867]: I1212 08:27:09.944396 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007550 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007637 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007683 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007742 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007773 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007854 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007887 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs6jf\" (UniqueName: \"kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007932 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6blt5\" (UniqueName: \"kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007953 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007977 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.007999 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.008033 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwbp6\" (UniqueName: \"kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.017024 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.022041 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.029938 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.051951 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs6jf\" (UniqueName: \"kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf\") pod \"heat-engine-87678c97b-9xgm2\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.107747 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.109409 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.109603 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.109710 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.109918 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.111760 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6blt5\" (UniqueName: \"kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.111933 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.112083 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwbp6\" (UniqueName: \"kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.115930 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.115940 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.117144 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.117383 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.133579 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.136195 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwbp6\" (UniqueName: \"kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6\") pod \"heat-cfnapi-cd4854554-7lbcd\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.144495 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.144472 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.153068 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6blt5\" (UniqueName: \"kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5\") pod \"heat-api-9f4d674c-9qtmz\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.193742 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.284738 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.711682 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.804128 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:10 crc kubenswrapper[4867]: I1212 08:27:10.926351 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:10 crc kubenswrapper[4867]: W1212 08:27:10.927373 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48e5d217_dd86_4762_a240_1d02cfd88a4c.slice/crio-677201a07bd36cdeac19db29da18f90779dc51d0bd4030670993d54be3ef2714 WatchSource:0}: Error finding container 677201a07bd36cdeac19db29da18f90779dc51d0bd4030670993d54be3ef2714: Status 404 returned error can't find the container with id 677201a07bd36cdeac19db29da18f90779dc51d0bd4030670993d54be3ef2714 Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.631731 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9f4d674c-9qtmz" event={"ID":"97e66252-f3dc-4d51-a515-4aee4cea9f5e","Type":"ContainerStarted","Data":"edb7cb492f706925b776be8073c2e56a31e43bede376b7633e52342711df08ff"} Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.634293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cd4854554-7lbcd" event={"ID":"48e5d217-dd86-4762-a240-1d02cfd88a4c","Type":"ContainerStarted","Data":"677201a07bd36cdeac19db29da18f90779dc51d0bd4030670993d54be3ef2714"} Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.636559 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-87678c97b-9xgm2" event={"ID":"988ad98d-bd39-451b-a946-c04cbe4f4f85","Type":"ContainerStarted","Data":"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1"} Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.636600 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-87678c97b-9xgm2" event={"ID":"988ad98d-bd39-451b-a946-c04cbe4f4f85","Type":"ContainerStarted","Data":"029ec3fec8c172192a9562778a85c6097b2d5a4ac7afa6e725abc4d8e067708f"} Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.636806 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:11 crc kubenswrapper[4867]: I1212 08:27:11.658945 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-87678c97b-9xgm2" podStartSLOduration=2.658921346 podStartE2EDuration="2.658921346s" podCreationTimestamp="2025-12-12 08:27:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:11.654940708 +0000 UTC m=+5919.226321977" watchObservedRunningTime="2025-12-12 08:27:11.658921346 +0000 UTC m=+5919.230302615" Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.656697 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9f4d674c-9qtmz" event={"ID":"97e66252-f3dc-4d51-a515-4aee4cea9f5e","Type":"ContainerStarted","Data":"e1d3299b7ebb14c9da89f21415f0547cfe791087609ad296543b6da9c0921854"} Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.657398 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.659186 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cd4854554-7lbcd" event={"ID":"48e5d217-dd86-4762-a240-1d02cfd88a4c","Type":"ContainerStarted","Data":"9e806eff9804b2351bbc905559efd88d4825e58cf326718e5356db3f2f72653a"} Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.659728 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.677682 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-9f4d674c-9qtmz" podStartSLOduration=2.227827803 podStartE2EDuration="4.677662717s" podCreationTimestamp="2025-12-12 08:27:09 +0000 UTC" firstStartedPulling="2025-12-12 08:27:10.811985999 +0000 UTC m=+5918.383367268" lastFinishedPulling="2025-12-12 08:27:13.261820913 +0000 UTC m=+5920.833202182" observedRunningTime="2025-12-12 08:27:13.676000516 +0000 UTC m=+5921.247381805" watchObservedRunningTime="2025-12-12 08:27:13.677662717 +0000 UTC m=+5921.249043986" Dec 12 08:27:13 crc kubenswrapper[4867]: I1212 08:27:13.702434 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-cd4854554-7lbcd" podStartSLOduration=2.363810037 podStartE2EDuration="4.702414481s" podCreationTimestamp="2025-12-12 08:27:09 +0000 UTC" firstStartedPulling="2025-12-12 08:27:10.929870044 +0000 UTC m=+5918.501251313" lastFinishedPulling="2025-12-12 08:27:13.268474488 +0000 UTC m=+5920.839855757" observedRunningTime="2025-12-12 08:27:13.693440318 +0000 UTC m=+5921.264821607" watchObservedRunningTime="2025-12-12 08:27:13.702414481 +0000 UTC m=+5921.273795750" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.485722 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6dfc547d9b-s2cdv"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.487999 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.500563 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.502176 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.526810 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.528314 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.547015 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.557797 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578702 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkxb9\" (UniqueName: \"kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578761 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578833 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-combined-ca-bundle\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578897 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578938 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data-custom\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.578970 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579093 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579139 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579172 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579213 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj5m8\" (UniqueName: \"kubernetes.io/projected/642472d4-e097-4004-bbd1-cbb5f11fc76d-kube-api-access-wj5m8\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.579361 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ll74\" (UniqueName: \"kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.589996 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6dfc547d9b-s2cdv"] Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.680912 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.680987 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681033 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681092 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj5m8\" (UniqueName: \"kubernetes.io/projected/642472d4-e097-4004-bbd1-cbb5f11fc76d-kube-api-access-wj5m8\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ll74\" (UniqueName: \"kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681177 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkxb9\" (UniqueName: \"kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681209 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681266 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-combined-ca-bundle\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681327 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681370 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data-custom\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681403 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.681445 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.690674 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.691450 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.691717 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.692213 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.692213 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-combined-ca-bundle\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.696512 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.697351 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.707857 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj5m8\" (UniqueName: \"kubernetes.io/projected/642472d4-e097-4004-bbd1-cbb5f11fc76d-kube-api-access-wj5m8\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.710619 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/642472d4-e097-4004-bbd1-cbb5f11fc76d-config-data-custom\") pod \"heat-engine-6dfc547d9b-s2cdv\" (UID: \"642472d4-e097-4004-bbd1-cbb5f11fc76d\") " pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.712763 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.713955 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ll74\" (UniqueName: \"kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74\") pod \"heat-cfnapi-59bb949fb8-p7xt5\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.714186 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkxb9\" (UniqueName: \"kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9\") pod \"heat-api-58487cbbc9-hkg9v\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.810909 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.819466 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:17 crc kubenswrapper[4867]: I1212 08:27:17.847310 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.372132 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.387171 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6dfc547d9b-s2cdv"] Dec 12 08:27:18 crc kubenswrapper[4867]: W1212 08:27:18.392938 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73deaf53_d7ca_4849_9e60_aea78ef12e0f.slice/crio-bc44d093892559f25114dcc825de84b05084d1a23447d27195440ed392782f8a WatchSource:0}: Error finding container bc44d093892559f25114dcc825de84b05084d1a23447d27195440ed392782f8a: Status 404 returned error can't find the container with id bc44d093892559f25114dcc825de84b05084d1a23447d27195440ed392782f8a Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.403645 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.615430 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.615931 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-9f4d674c-9qtmz" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" containerID="cri-o://e1d3299b7ebb14c9da89f21415f0547cfe791087609ad296543b6da9c0921854" gracePeriod=60 Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.641570 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-9f4d674c-9qtmz" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.108:8004/healthcheck\": EOF" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.642535 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.642777 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-cd4854554-7lbcd" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerName="heat-cfnapi" containerID="cri-o://9e806eff9804b2351bbc905559efd88d4825e58cf326718e5356db3f2f72653a" gracePeriod=60 Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.656954 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7b64b9cdf4-96grl"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.658112 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.662896 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.663490 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.677887 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b64b9cdf4-96grl"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700447 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700583 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z5qh\" (UniqueName: \"kubernetes.io/projected/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-kube-api-access-6z5qh\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700626 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-internal-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700664 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-public-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700722 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data-custom\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.700750 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-combined-ca-bundle\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.729780 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-864798f469-5rmb7"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.731840 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.735286 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.735598 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.766674 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-58487cbbc9-hkg9v" event={"ID":"ac9cedf1-6026-4e13-9323-9bd9f81006e8","Type":"ContainerStarted","Data":"4a66f356c2dc22e203b62c19be9d7470117bb6102d8920bec07f785c837978d9"} Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.766809 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-58487cbbc9-hkg9v" event={"ID":"ac9cedf1-6026-4e13-9323-9bd9f81006e8","Type":"ContainerStarted","Data":"1c10057d529e6671d1577c83153cfc3a374c52b8df7f3cd3c9ab575c457feca5"} Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.767503 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.797034 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-864798f469-5rmb7"] Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.797077 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" event={"ID":"73deaf53-d7ca-4849-9e60-aea78ef12e0f","Type":"ContainerStarted","Data":"bc44d093892559f25114dcc825de84b05084d1a23447d27195440ed392782f8a"} Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.804166 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.821109 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfc547d9b-s2cdv" event={"ID":"642472d4-e097-4004-bbd1-cbb5f11fc76d","Type":"ContainerStarted","Data":"d0283d2cfb34ae0ba36d612660b7e49928254a45eb79becb088601834986cbcb"} Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.821433 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.836261 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wzs\" (UniqueName: \"kubernetes.io/projected/975458e4-1347-4cd8-8391-a7a95dfda548-kube-api-access-j9wzs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.836433 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z5qh\" (UniqueName: \"kubernetes.io/projected/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-kube-api-access-6z5qh\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837388 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-combined-ca-bundle\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837570 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-internal-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837667 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-public-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837752 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-public-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837803 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data-custom\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837913 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data-custom\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.837984 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-combined-ca-bundle\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.838034 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.838103 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.838143 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-internal-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.843771 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-combined-ca-bundle\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.864025 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.865067 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-public-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.869068 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-58487cbbc9-hkg9v" podStartSLOduration=1.8690440879999999 podStartE2EDuration="1.869044088s" podCreationTimestamp="2025-12-12 08:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:18.811313296 +0000 UTC m=+5926.382694565" watchObservedRunningTime="2025-12-12 08:27:18.869044088 +0000 UTC m=+5926.440425357" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.869838 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-config-data-custom\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.878440 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" podStartSLOduration=1.878419241 podStartE2EDuration="1.878419241s" podCreationTimestamp="2025-12-12 08:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:18.866076754 +0000 UTC m=+5926.437458023" watchObservedRunningTime="2025-12-12 08:27:18.878419241 +0000 UTC m=+5926.449800510" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.885340 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z5qh\" (UniqueName: \"kubernetes.io/projected/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-kube-api-access-6z5qh\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.885665 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e214fa4-0f89-41fa-8e9b-d064adcd4c3c-internal-tls-certs\") pod \"heat-api-7b64b9cdf4-96grl\" (UID: \"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c\") " pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.887356 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6dfc547d9b-s2cdv" podStartSLOduration=1.887342512 podStartE2EDuration="1.887342512s" podCreationTimestamp="2025-12-12 08:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:18.883760942 +0000 UTC m=+5926.455142211" watchObservedRunningTime="2025-12-12 08:27:18.887342512 +0000 UTC m=+5926.458723781" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.940558 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-public-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.940618 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data-custom\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.940805 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.940832 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-internal-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.940948 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wzs\" (UniqueName: \"kubernetes.io/projected/975458e4-1347-4cd8-8391-a7a95dfda548-kube-api-access-j9wzs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.941033 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-combined-ca-bundle\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.947192 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-public-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.948210 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-combined-ca-bundle\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.951618 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.951753 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-config-data-custom\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.953958 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/975458e4-1347-4cd8-8391-a7a95dfda548-internal-tls-certs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:18 crc kubenswrapper[4867]: I1212 08:27:18.964166 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wzs\" (UniqueName: \"kubernetes.io/projected/975458e4-1347-4cd8-8391-a7a95dfda548-kube-api-access-j9wzs\") pod \"heat-cfnapi-864798f469-5rmb7\" (UID: \"975458e4-1347-4cd8-8391-a7a95dfda548\") " pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.067085 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.099454 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:19 crc kubenswrapper[4867]: W1212 08:27:19.620385 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e214fa4_0f89_41fa_8e9b_d064adcd4c3c.slice/crio-a3ad915885ef543308145f4726cd0f191f249e0914e3023de1d8d0bb6445e293 WatchSource:0}: Error finding container a3ad915885ef543308145f4726cd0f191f249e0914e3023de1d8d0bb6445e293: Status 404 returned error can't find the container with id a3ad915885ef543308145f4726cd0f191f249e0914e3023de1d8d0bb6445e293 Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.620411 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b64b9cdf4-96grl"] Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.737422 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-864798f469-5rmb7"] Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.851205 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b64b9cdf4-96grl" event={"ID":"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c","Type":"ContainerStarted","Data":"a3ad915885ef543308145f4726cd0f191f249e0914e3023de1d8d0bb6445e293"} Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.854299 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864798f469-5rmb7" event={"ID":"975458e4-1347-4cd8-8391-a7a95dfda548","Type":"ContainerStarted","Data":"9246f2837d72f690d7fddfa731eb8b855ded085bce6dec4c4d8ebabe8563ed96"} Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.869800 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6dfc547d9b-s2cdv" event={"ID":"642472d4-e097-4004-bbd1-cbb5f11fc76d","Type":"ContainerStarted","Data":"11ba329415c466afc46c2363a86f94f7ba3ee7fe1519eee79b8f4466144f9ef0"} Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.877731 4867 generic.go:334] "Generic (PLEG): container finished" podID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerID="96739c6d3865edf68ab3dadeb8ebb6d83d8f46aa175ded02aa511e875b5ef73a" exitCode=1 Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.877814 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" event={"ID":"73deaf53-d7ca-4849-9e60-aea78ef12e0f","Type":"ContainerDied","Data":"96739c6d3865edf68ab3dadeb8ebb6d83d8f46aa175ded02aa511e875b5ef73a"} Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.879427 4867 scope.go:117] "RemoveContainer" containerID="96739c6d3865edf68ab3dadeb8ebb6d83d8f46aa175ded02aa511e875b5ef73a" Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.886382 4867 generic.go:334] "Generic (PLEG): container finished" podID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerID="4a66f356c2dc22e203b62c19be9d7470117bb6102d8920bec07f785c837978d9" exitCode=1 Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.886427 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-58487cbbc9-hkg9v" event={"ID":"ac9cedf1-6026-4e13-9323-9bd9f81006e8","Type":"ContainerDied","Data":"4a66f356c2dc22e203b62c19be9d7470117bb6102d8920bec07f785c837978d9"} Dec 12 08:27:19 crc kubenswrapper[4867]: I1212 08:27:19.887063 4867 scope.go:117] "RemoveContainer" containerID="4a66f356c2dc22e203b62c19be9d7470117bb6102d8920bec07f785c837978d9" Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.206965 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.649246 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.933492 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-864798f469-5rmb7" event={"ID":"975458e4-1347-4cd8-8391-a7a95dfda548","Type":"ContainerStarted","Data":"4ac7da7d224ae50dbbc18a0ec947c401cd8d5305d9910194c23942cc747e5f0f"} Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.934742 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.970898 4867 generic.go:334] "Generic (PLEG): container finished" podID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerID="ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5" exitCode=1 Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.970997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-58487cbbc9-hkg9v" event={"ID":"ac9cedf1-6026-4e13-9323-9bd9f81006e8","Type":"ContainerDied","Data":"ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5"} Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.971037 4867 scope.go:117] "RemoveContainer" containerID="4a66f356c2dc22e203b62c19be9d7470117bb6102d8920bec07f785c837978d9" Dec 12 08:27:20 crc kubenswrapper[4867]: I1212 08:27:20.971806 4867 scope.go:117] "RemoveContainer" containerID="ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5" Dec 12 08:27:20 crc kubenswrapper[4867]: E1212 08:27:20.972067 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-58487cbbc9-hkg9v_openstack(ac9cedf1-6026-4e13-9323-9bd9f81006e8)\"" pod="openstack/heat-api-58487cbbc9-hkg9v" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.011509 4867 generic.go:334] "Generic (PLEG): container finished" podID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerID="1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8" exitCode=1 Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.011905 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" event={"ID":"73deaf53-d7ca-4849-9e60-aea78ef12e0f","Type":"ContainerDied","Data":"1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8"} Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.012602 4867 scope.go:117] "RemoveContainer" containerID="1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8" Dec 12 08:27:21 crc kubenswrapper[4867]: E1212 08:27:21.012866 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59bb949fb8-p7xt5_openstack(73deaf53-d7ca-4849-9e60-aea78ef12e0f)\"" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.053886 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-864798f469-5rmb7" podStartSLOduration=3.053859537 podStartE2EDuration="3.053859537s" podCreationTimestamp="2025-12-12 08:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:20.98060507 +0000 UTC m=+5928.551986339" watchObservedRunningTime="2025-12-12 08:27:21.053859537 +0000 UTC m=+5928.625240806" Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.114374 4867 scope.go:117] "RemoveContainer" containerID="96739c6d3865edf68ab3dadeb8ebb6d83d8f46aa175ded02aa511e875b5ef73a" Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.134823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b64b9cdf4-96grl" event={"ID":"1e214fa4-0f89-41fa-8e9b-d064adcd4c3c","Type":"ContainerStarted","Data":"66140ae68c64cff25e8ac9a1387bdaf7ba960be20aa3f6963bcb2720a3f844ad"} Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.134972 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:21 crc kubenswrapper[4867]: I1212 08:27:21.188050 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7b64b9cdf4-96grl" podStartSLOduration=3.188027655 podStartE2EDuration="3.188027655s" podCreationTimestamp="2025-12-12 08:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:27:21.168786747 +0000 UTC m=+5928.740168016" watchObservedRunningTime="2025-12-12 08:27:21.188027655 +0000 UTC m=+5928.759408924" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.126650 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.151271 4867 scope.go:117] "RemoveContainer" containerID="ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5" Dec 12 08:27:22 crc kubenswrapper[4867]: E1212 08:27:22.151520 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-58487cbbc9-hkg9v_openstack(ac9cedf1-6026-4e13-9323-9bd9f81006e8)\"" pod="openstack/heat-api-58487cbbc9-hkg9v" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.159711 4867 scope.go:117] "RemoveContainer" containerID="1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8" Dec 12 08:27:22 crc kubenswrapper[4867]: E1212 08:27:22.159940 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59bb949fb8-p7xt5_openstack(73deaf53-d7ca-4849-9e60-aea78ef12e0f)\"" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.820413 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.820808 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.856500 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:22 crc kubenswrapper[4867]: I1212 08:27:22.856541 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:23 crc kubenswrapper[4867]: I1212 08:27:23.182350 4867 scope.go:117] "RemoveContainer" containerID="1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8" Dec 12 08:27:23 crc kubenswrapper[4867]: E1212 08:27:23.182924 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-59bb949fb8-p7xt5_openstack(73deaf53-d7ca-4849-9e60-aea78ef12e0f)\"" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" Dec 12 08:27:23 crc kubenswrapper[4867]: I1212 08:27:23.183262 4867 scope.go:117] "RemoveContainer" containerID="ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5" Dec 12 08:27:23 crc kubenswrapper[4867]: E1212 08:27:23.183671 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-58487cbbc9-hkg9v_openstack(ac9cedf1-6026-4e13-9323-9bd9f81006e8)\"" pod="openstack/heat-api-58487cbbc9-hkg9v" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.088670 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-cd4854554-7lbcd" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.109:8000/healthcheck\": read tcp 10.217.0.2:52690->10.217.1.109:8000: read: connection reset by peer" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.090685 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-9f4d674c-9qtmz" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.108:8004/healthcheck\": read tcp 10.217.0.2:46710->10.217.1.108:8004: read: connection reset by peer" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.192024 4867 generic.go:334] "Generic (PLEG): container finished" podID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerID="e1d3299b7ebb14c9da89f21415f0547cfe791087609ad296543b6da9c0921854" exitCode=0 Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.192395 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9f4d674c-9qtmz" event={"ID":"97e66252-f3dc-4d51-a515-4aee4cea9f5e","Type":"ContainerDied","Data":"e1d3299b7ebb14c9da89f21415f0547cfe791087609ad296543b6da9c0921854"} Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.195724 4867 generic.go:334] "Generic (PLEG): container finished" podID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerID="9e806eff9804b2351bbc905559efd88d4825e58cf326718e5356db3f2f72653a" exitCode=0 Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.195763 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cd4854554-7lbcd" event={"ID":"48e5d217-dd86-4762-a240-1d02cfd88a4c","Type":"ContainerDied","Data":"9e806eff9804b2351bbc905559efd88d4825e58cf326718e5356db3f2f72653a"} Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.203731 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f8df8656d-5rfkc" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.301900 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.302528 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" containerID="cri-o://0c2ef26e9e0539234a2711588629f1b4c3192e1df0090da32f310134726c47a6" gracePeriod=30 Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.312183 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon-log" containerID="cri-o://65f20d52bcd69dd154a5e3248e97cb73283ebb3f00a7bb4fdbb759a105e3048d" gracePeriod=30 Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.835654 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.848061 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.916640 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom\") pod \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.916744 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwbp6\" (UniqueName: \"kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6\") pod \"48e5d217-dd86-4762-a240-1d02cfd88a4c\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.918572 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data\") pod \"48e5d217-dd86-4762-a240-1d02cfd88a4c\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.918685 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data\") pod \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.918801 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle\") pod \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.918883 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6blt5\" (UniqueName: \"kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5\") pod \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\" (UID: \"97e66252-f3dc-4d51-a515-4aee4cea9f5e\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.918928 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom\") pod \"48e5d217-dd86-4762-a240-1d02cfd88a4c\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.919015 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle\") pod \"48e5d217-dd86-4762-a240-1d02cfd88a4c\" (UID: \"48e5d217-dd86-4762-a240-1d02cfd88a4c\") " Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.925143 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6" (OuterVolumeSpecName: "kube-api-access-mwbp6") pod "48e5d217-dd86-4762-a240-1d02cfd88a4c" (UID: "48e5d217-dd86-4762-a240-1d02cfd88a4c"). InnerVolumeSpecName "kube-api-access-mwbp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.925396 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5" (OuterVolumeSpecName: "kube-api-access-6blt5") pod "97e66252-f3dc-4d51-a515-4aee4cea9f5e" (UID: "97e66252-f3dc-4d51-a515-4aee4cea9f5e"). InnerVolumeSpecName "kube-api-access-6blt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.926683 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "48e5d217-dd86-4762-a240-1d02cfd88a4c" (UID: "48e5d217-dd86-4762-a240-1d02cfd88a4c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.931100 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "97e66252-f3dc-4d51-a515-4aee4cea9f5e" (UID: "97e66252-f3dc-4d51-a515-4aee4cea9f5e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.957803 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97e66252-f3dc-4d51-a515-4aee4cea9f5e" (UID: "97e66252-f3dc-4d51-a515-4aee4cea9f5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.966424 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48e5d217-dd86-4762-a240-1d02cfd88a4c" (UID: "48e5d217-dd86-4762-a240-1d02cfd88a4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:24 crc kubenswrapper[4867]: I1212 08:27:24.993988 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data" (OuterVolumeSpecName: "config-data") pod "97e66252-f3dc-4d51-a515-4aee4cea9f5e" (UID: "97e66252-f3dc-4d51-a515-4aee4cea9f5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.019131 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data" (OuterVolumeSpecName: "config-data") pod "48e5d217-dd86-4762-a240-1d02cfd88a4c" (UID: "48e5d217-dd86-4762-a240-1d02cfd88a4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022486 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022529 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022544 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6blt5\" (UniqueName: \"kubernetes.io/projected/97e66252-f3dc-4d51-a515-4aee4cea9f5e-kube-api-access-6blt5\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022556 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022692 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022703 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97e66252-f3dc-4d51-a515-4aee4cea9f5e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022714 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwbp6\" (UniqueName: \"kubernetes.io/projected/48e5d217-dd86-4762-a240-1d02cfd88a4c-kube-api-access-mwbp6\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.022725 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48e5d217-dd86-4762-a240-1d02cfd88a4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.205870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-9f4d674c-9qtmz" event={"ID":"97e66252-f3dc-4d51-a515-4aee4cea9f5e","Type":"ContainerDied","Data":"edb7cb492f706925b776be8073c2e56a31e43bede376b7633e52342711df08ff"} Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.205903 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-9f4d674c-9qtmz" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.205934 4867 scope.go:117] "RemoveContainer" containerID="e1d3299b7ebb14c9da89f21415f0547cfe791087609ad296543b6da9c0921854" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.207724 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cd4854554-7lbcd" event={"ID":"48e5d217-dd86-4762-a240-1d02cfd88a4c","Type":"ContainerDied","Data":"677201a07bd36cdeac19db29da18f90779dc51d0bd4030670993d54be3ef2714"} Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.207810 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cd4854554-7lbcd" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.247436 4867 scope.go:117] "RemoveContainer" containerID="9e806eff9804b2351bbc905559efd88d4825e58cf326718e5356db3f2f72653a" Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.250662 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.260872 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-9f4d674c-9qtmz"] Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.270454 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:25 crc kubenswrapper[4867]: I1212 08:27:25.278620 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-cd4854554-7lbcd"] Dec 12 08:27:26 crc kubenswrapper[4867]: I1212 08:27:26.849122 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" path="/var/lib/kubelet/pods/48e5d217-dd86-4762-a240-1d02cfd88a4c/volumes" Dec 12 08:27:26 crc kubenswrapper[4867]: I1212 08:27:26.850091 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" path="/var/lib/kubelet/pods/97e66252-f3dc-4d51-a515-4aee4cea9f5e/volumes" Dec 12 08:27:27 crc kubenswrapper[4867]: I1212 08:27:27.842637 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.99:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:52164->10.217.1.99:8443: read: connection reset by peer" Dec 12 08:27:28 crc kubenswrapper[4867]: I1212 08:27:28.241211 4867 generic.go:334] "Generic (PLEG): container finished" podID="a52ae013-bec4-4e49-8322-9f291818039d" containerID="0c2ef26e9e0539234a2711588629f1b4c3192e1df0090da32f310134726c47a6" exitCode=0 Dec 12 08:27:28 crc kubenswrapper[4867]: I1212 08:27:28.241277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerDied","Data":"0c2ef26e9e0539234a2711588629f1b4c3192e1df0090da32f310134726c47a6"} Dec 12 08:27:28 crc kubenswrapper[4867]: I1212 08:27:28.989009 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:27:28 crc kubenswrapper[4867]: I1212 08:27:28.989376 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:27:30 crc kubenswrapper[4867]: I1212 08:27:30.416768 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7b64b9cdf4-96grl" Dec 12 08:27:30 crc kubenswrapper[4867]: I1212 08:27:30.485665 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-864798f469-5rmb7" Dec 12 08:27:30 crc kubenswrapper[4867]: I1212 08:27:30.492473 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:30 crc kubenswrapper[4867]: I1212 08:27:30.577338 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.046021 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.051079 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.150928 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkxb9\" (UniqueName: \"kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9\") pod \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151298 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom\") pod \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151467 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom\") pod \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151667 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data\") pod \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151773 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle\") pod \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151868 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle\") pod \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.151997 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data\") pod \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\" (UID: \"ac9cedf1-6026-4e13-9323-9bd9f81006e8\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.152123 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ll74\" (UniqueName: \"kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74\") pod \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\" (UID: \"73deaf53-d7ca-4849-9e60-aea78ef12e0f\") " Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.156802 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ac9cedf1-6026-4e13-9323-9bd9f81006e8" (UID: "ac9cedf1-6026-4e13-9323-9bd9f81006e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.156829 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "73deaf53-d7ca-4849-9e60-aea78ef12e0f" (UID: "73deaf53-d7ca-4849-9e60-aea78ef12e0f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.157216 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74" (OuterVolumeSpecName: "kube-api-access-7ll74") pod "73deaf53-d7ca-4849-9e60-aea78ef12e0f" (UID: "73deaf53-d7ca-4849-9e60-aea78ef12e0f"). InnerVolumeSpecName "kube-api-access-7ll74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.157405 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9" (OuterVolumeSpecName: "kube-api-access-lkxb9") pod "ac9cedf1-6026-4e13-9323-9bd9f81006e8" (UID: "ac9cedf1-6026-4e13-9323-9bd9f81006e8"). InnerVolumeSpecName "kube-api-access-lkxb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.184474 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73deaf53-d7ca-4849-9e60-aea78ef12e0f" (UID: "73deaf53-d7ca-4849-9e60-aea78ef12e0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.188559 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac9cedf1-6026-4e13-9323-9bd9f81006e8" (UID: "ac9cedf1-6026-4e13-9323-9bd9f81006e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.211260 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data" (OuterVolumeSpecName: "config-data") pod "ac9cedf1-6026-4e13-9323-9bd9f81006e8" (UID: "ac9cedf1-6026-4e13-9323-9bd9f81006e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.221462 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data" (OuterVolumeSpecName: "config-data") pod "73deaf53-d7ca-4849-9e60-aea78ef12e0f" (UID: "73deaf53-d7ca-4849-9e60-aea78ef12e0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255448 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255482 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255492 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255501 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73deaf53-d7ca-4849-9e60-aea78ef12e0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255509 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255518 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac9cedf1-6026-4e13-9323-9bd9f81006e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255528 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ll74\" (UniqueName: \"kubernetes.io/projected/73deaf53-d7ca-4849-9e60-aea78ef12e0f-kube-api-access-7ll74\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.255540 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkxb9\" (UniqueName: \"kubernetes.io/projected/ac9cedf1-6026-4e13-9323-9bd9f81006e8-kube-api-access-lkxb9\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.272204 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-58487cbbc9-hkg9v" event={"ID":"ac9cedf1-6026-4e13-9323-9bd9f81006e8","Type":"ContainerDied","Data":"1c10057d529e6671d1577c83153cfc3a374c52b8df7f3cd3c9ab575c457feca5"} Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.272260 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-58487cbbc9-hkg9v" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.272331 4867 scope.go:117] "RemoveContainer" containerID="ae5a18a25bcdd73784324138f6faaf62fbc275ba1fd83de967a5d0e977fd06c5" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.281094 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" event={"ID":"73deaf53-d7ca-4849-9e60-aea78ef12e0f","Type":"ContainerDied","Data":"bc44d093892559f25114dcc825de84b05084d1a23447d27195440ed392782f8a"} Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.281189 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-59bb949fb8-p7xt5" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.335459 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.345414 4867 scope.go:117] "RemoveContainer" containerID="1d80bd26aa3e4c2e0141425630c4cae3d7f6599a74ac34b09cd4a7b040f91de8" Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.350392 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-59bb949fb8-p7xt5"] Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.359962 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:31 crc kubenswrapper[4867]: I1212 08:27:31.371818 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-58487cbbc9-hkg9v"] Dec 12 08:27:32 crc kubenswrapper[4867]: I1212 08:27:32.855805 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" path="/var/lib/kubelet/pods/73deaf53-d7ca-4849-9e60-aea78ef12e0f/volumes" Dec 12 08:27:32 crc kubenswrapper[4867]: I1212 08:27:32.857306 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" path="/var/lib/kubelet/pods/ac9cedf1-6026-4e13-9323-9bd9f81006e8/volumes" Dec 12 08:27:34 crc kubenswrapper[4867]: I1212 08:27:34.720868 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.99:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.99:8443: connect: connection refused" Dec 12 08:27:37 crc kubenswrapper[4867]: I1212 08:27:37.843482 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6dfc547d9b-s2cdv" Dec 12 08:27:37 crc kubenswrapper[4867]: I1212 08:27:37.892110 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:37 crc kubenswrapper[4867]: I1212 08:27:37.893077 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-87678c97b-9xgm2" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" containerID="cri-o://20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" gracePeriod=60 Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592048 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592614 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592638 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592659 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592667 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592685 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592694 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592710 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592718 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592737 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592744 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: E1212 08:27:38.592765 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592771 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592949 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="97e66252-f3dc-4d51-a515-4aee4cea9f5e" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592957 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592965 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="73deaf53-d7ca-4849-9e60-aea78ef12e0f" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592975 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.592990 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac9cedf1-6026-4e13-9323-9bd9f81006e8" containerName="heat-api" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.593000 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e5d217-dd86-4762-a240-1d02cfd88a4c" containerName="heat-cfnapi" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.594428 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.597481 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.597529 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkxt\" (UniqueName: \"kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.597690 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.604804 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.700346 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.700619 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mkxt\" (UniqueName: \"kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.700804 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.701289 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.701456 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.729033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mkxt\" (UniqueName: \"kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt\") pod \"redhat-operators-h65wz\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:38 crc kubenswrapper[4867]: I1212 08:27:38.925199 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:39 crc kubenswrapper[4867]: I1212 08:27:39.465574 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:27:40 crc kubenswrapper[4867]: E1212 08:27:40.110360 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:40 crc kubenswrapper[4867]: E1212 08:27:40.112763 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:40 crc kubenswrapper[4867]: E1212 08:27:40.114840 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:40 crc kubenswrapper[4867]: E1212 08:27:40.114895 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-87678c97b-9xgm2" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" Dec 12 08:27:40 crc kubenswrapper[4867]: I1212 08:27:40.376784 4867 generic.go:334] "Generic (PLEG): container finished" podID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerID="ac3f2ddd560a04048992597f7203994f6cbf5fa3f6d152da5b883bb8e9c25e52" exitCode=0 Dec 12 08:27:40 crc kubenswrapper[4867]: I1212 08:27:40.376859 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerDied","Data":"ac3f2ddd560a04048992597f7203994f6cbf5fa3f6d152da5b883bb8e9c25e52"} Dec 12 08:27:40 crc kubenswrapper[4867]: I1212 08:27:40.376908 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerStarted","Data":"194ee2518bc08a697f64fc1dab74460e96ec61fb71f91283304029d6fa9edd97"} Dec 12 08:27:42 crc kubenswrapper[4867]: I1212 08:27:42.396621 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerStarted","Data":"b890f658fc9de0b0ff0feedd3b93577c57000156fed9605aefb5775f7b886337"} Dec 12 08:27:44 crc kubenswrapper[4867]: I1212 08:27:44.416377 4867 generic.go:334] "Generic (PLEG): container finished" podID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerID="b890f658fc9de0b0ff0feedd3b93577c57000156fed9605aefb5775f7b886337" exitCode=0 Dec 12 08:27:44 crc kubenswrapper[4867]: I1212 08:27:44.416495 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerDied","Data":"b890f658fc9de0b0ff0feedd3b93577c57000156fed9605aefb5775f7b886337"} Dec 12 08:27:44 crc kubenswrapper[4867]: I1212 08:27:44.720369 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.99:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.99:8443: connect: connection refused" Dec 12 08:27:44 crc kubenswrapper[4867]: I1212 08:27:44.720935 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:27:46 crc kubenswrapper[4867]: I1212 08:27:46.443526 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerStarted","Data":"f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006"} Dec 12 08:27:46 crc kubenswrapper[4867]: I1212 08:27:46.478665 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h65wz" podStartSLOduration=3.399627261 podStartE2EDuration="8.478628544s" podCreationTimestamp="2025-12-12 08:27:38 +0000 UTC" firstStartedPulling="2025-12-12 08:27:40.381621712 +0000 UTC m=+5947.953002981" lastFinishedPulling="2025-12-12 08:27:45.460622995 +0000 UTC m=+5953.032004264" observedRunningTime="2025-12-12 08:27:46.467109339 +0000 UTC m=+5954.038490628" watchObservedRunningTime="2025-12-12 08:27:46.478628544 +0000 UTC m=+5954.050009813" Dec 12 08:27:48 crc kubenswrapper[4867]: I1212 08:27:48.926119 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:48 crc kubenswrapper[4867]: I1212 08:27:48.926477 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:49 crc kubenswrapper[4867]: I1212 08:27:49.974221 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h65wz" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="registry-server" probeResult="failure" output=< Dec 12 08:27:49 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 08:27:49 crc kubenswrapper[4867]: > Dec 12 08:27:50 crc kubenswrapper[4867]: E1212 08:27:50.111106 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 is running failed: container process not found" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:50 crc kubenswrapper[4867]: E1212 08:27:50.111685 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 is running failed: container process not found" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:50 crc kubenswrapper[4867]: E1212 08:27:50.111950 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 is running failed: container process not found" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 12 08:27:50 crc kubenswrapper[4867]: E1212 08:27:50.111983 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 is running failed: container process not found" probeType="Readiness" pod="openstack/heat-engine-87678c97b-9xgm2" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.349282 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.473440 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data\") pod \"988ad98d-bd39-451b-a946-c04cbe4f4f85\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.473743 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs6jf\" (UniqueName: \"kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf\") pod \"988ad98d-bd39-451b-a946-c04cbe4f4f85\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.473944 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle\") pod \"988ad98d-bd39-451b-a946-c04cbe4f4f85\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.474098 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom\") pod \"988ad98d-bd39-451b-a946-c04cbe4f4f85\" (UID: \"988ad98d-bd39-451b-a946-c04cbe4f4f85\") " Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.479074 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf" (OuterVolumeSpecName: "kube-api-access-cs6jf") pod "988ad98d-bd39-451b-a946-c04cbe4f4f85" (UID: "988ad98d-bd39-451b-a946-c04cbe4f4f85"). InnerVolumeSpecName "kube-api-access-cs6jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.479410 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "988ad98d-bd39-451b-a946-c04cbe4f4f85" (UID: "988ad98d-bd39-451b-a946-c04cbe4f4f85"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.485992 4867 generic.go:334] "Generic (PLEG): container finished" podID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" exitCode=0 Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.486047 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-87678c97b-9xgm2" event={"ID":"988ad98d-bd39-451b-a946-c04cbe4f4f85","Type":"ContainerDied","Data":"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1"} Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.486081 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-87678c97b-9xgm2" event={"ID":"988ad98d-bd39-451b-a946-c04cbe4f4f85","Type":"ContainerDied","Data":"029ec3fec8c172192a9562778a85c6097b2d5a4ac7afa6e725abc4d8e067708f"} Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.486099 4867 scope.go:117] "RemoveContainer" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.486301 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-87678c97b-9xgm2" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.513346 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "988ad98d-bd39-451b-a946-c04cbe4f4f85" (UID: "988ad98d-bd39-451b-a946-c04cbe4f4f85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.541586 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data" (OuterVolumeSpecName: "config-data") pod "988ad98d-bd39-451b-a946-c04cbe4f4f85" (UID: "988ad98d-bd39-451b-a946-c04cbe4f4f85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.576995 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.577040 4867 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.577051 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/988ad98d-bd39-451b-a946-c04cbe4f4f85-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.577061 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs6jf\" (UniqueName: \"kubernetes.io/projected/988ad98d-bd39-451b-a946-c04cbe4f4f85-kube-api-access-cs6jf\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.588881 4867 scope.go:117] "RemoveContainer" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" Dec 12 08:27:50 crc kubenswrapper[4867]: E1212 08:27:50.589369 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1\": container with ID starting with 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 not found: ID does not exist" containerID="20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.589401 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1"} err="failed to get container status \"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1\": rpc error: code = NotFound desc = could not find container \"20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1\": container with ID starting with 20ed1edddc988206d651827e4631b3972991aad729b0a03bc0fec24e53964ff1 not found: ID does not exist" Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.818891 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.829628 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-87678c97b-9xgm2"] Dec 12 08:27:50 crc kubenswrapper[4867]: I1212 08:27:50.848268 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" path="/var/lib/kubelet/pods/988ad98d-bd39-451b-a946-c04cbe4f4f85/volumes" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.638116 4867 generic.go:334] "Generic (PLEG): container finished" podID="a52ae013-bec4-4e49-8322-9f291818039d" containerID="65f20d52bcd69dd154a5e3248e97cb73283ebb3f00a7bb4fdbb759a105e3048d" exitCode=137 Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.638173 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerDied","Data":"65f20d52bcd69dd154a5e3248e97cb73283ebb3f00a7bb4fdbb759a105e3048d"} Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.769505 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.874994 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.875361 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5sk7\" (UniqueName: \"kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.876017 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.876105 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.876335 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.876483 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.876599 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key\") pod \"a52ae013-bec4-4e49-8322-9f291818039d\" (UID: \"a52ae013-bec4-4e49-8322-9f291818039d\") " Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.877271 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs" (OuterVolumeSpecName: "logs") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.882486 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7" (OuterVolumeSpecName: "kube-api-access-m5sk7") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "kube-api-access-m5sk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.883363 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.909531 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.916865 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data" (OuterVolumeSpecName: "config-data") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.926285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts" (OuterVolumeSpecName: "scripts") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.940541 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a52ae013-bec4-4e49-8322-9f291818039d" (UID: "a52ae013-bec4-4e49-8322-9f291818039d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979720 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a52ae013-bec4-4e49-8322-9f291818039d-logs\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979878 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979941 4867 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979956 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a52ae013-bec4-4e49-8322-9f291818039d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979967 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5sk7\" (UniqueName: \"kubernetes.io/projected/a52ae013-bec4-4e49-8322-9f291818039d-kube-api-access-m5sk7\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979982 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:54 crc kubenswrapper[4867]: I1212 08:27:54.979993 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a52ae013-bec4-4e49-8322-9f291818039d-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.170871 4867 scope.go:117] "RemoveContainer" containerID="4e3809d01fc107522673b9e7d3500c16fe4c754545c20db485219df3a10ab2bc" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.610792 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th"] Dec 12 08:27:55 crc kubenswrapper[4867]: E1212 08:27:55.611593 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611617 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" Dec 12 08:27:55 crc kubenswrapper[4867]: E1212 08:27:55.611644 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611652 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" Dec 12 08:27:55 crc kubenswrapper[4867]: E1212 08:27:55.611671 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon-log" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611677 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon-log" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611876 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="988ad98d-bd39-451b-a946-c04cbe4f4f85" containerName="heat-engine" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611899 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon-log" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.611913 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.613685 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.615933 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.621357 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th"] Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.649929 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d5d9f84d-gcsp7" event={"ID":"a52ae013-bec4-4e49-8322-9f291818039d","Type":"ContainerDied","Data":"1bc920922ffd5db408450f400a39510688b651cc9a0570d439e0dbca5e592dc4"} Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.649988 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d5d9f84d-gcsp7" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.649997 4867 scope.go:117] "RemoveContainer" containerID="0c2ef26e9e0539234a2711588629f1b4c3192e1df0090da32f310134726c47a6" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.694576 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.694666 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q58sw\" (UniqueName: \"kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.694700 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.704377 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.716321 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d5d9f84d-gcsp7"] Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.797114 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.797219 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q58sw\" (UniqueName: \"kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.797285 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.797631 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.797669 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.818601 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q58sw\" (UniqueName: \"kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.839217 4867 scope.go:117] "RemoveContainer" containerID="65f20d52bcd69dd154a5e3248e97cb73283ebb3f00a7bb4fdbb759a105e3048d" Dec 12 08:27:55 crc kubenswrapper[4867]: I1212 08:27:55.986353 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:27:56 crc kubenswrapper[4867]: I1212 08:27:56.433945 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th"] Dec 12 08:27:56 crc kubenswrapper[4867]: I1212 08:27:56.678102 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerStarted","Data":"8610d59629b5323f1f8440f9a58e06c5bc46cd5a5d57fbe0d5167dc6bea8e3d1"} Dec 12 08:27:56 crc kubenswrapper[4867]: I1212 08:27:56.678472 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerStarted","Data":"83f1aaa0e2d88cd3de9758ecb82af2baeba1bbfa440a6466dd2cfdca79a60c15"} Dec 12 08:27:56 crc kubenswrapper[4867]: I1212 08:27:56.849748 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52ae013-bec4-4e49-8322-9f291818039d" path="/var/lib/kubelet/pods/a52ae013-bec4-4e49-8322-9f291818039d/volumes" Dec 12 08:27:57 crc kubenswrapper[4867]: I1212 08:27:57.688853 4867 generic.go:334] "Generic (PLEG): container finished" podID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerID="8610d59629b5323f1f8440f9a58e06c5bc46cd5a5d57fbe0d5167dc6bea8e3d1" exitCode=0 Dec 12 08:27:57 crc kubenswrapper[4867]: I1212 08:27:57.688891 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerDied","Data":"8610d59629b5323f1f8440f9a58e06c5bc46cd5a5d57fbe0d5167dc6bea8e3d1"} Dec 12 08:27:58 crc kubenswrapper[4867]: I1212 08:27:58.981202 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:58 crc kubenswrapper[4867]: I1212 08:27:58.989304 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:27:58 crc kubenswrapper[4867]: I1212 08:27:58.989347 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:27:59 crc kubenswrapper[4867]: I1212 08:27:59.045538 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:27:59 crc kubenswrapper[4867]: I1212 08:27:59.706968 4867 generic.go:334] "Generic (PLEG): container finished" podID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerID="d1d728f65936f38fd0eb61a1d97ab7f7117070ec9ef659662b4bb39409baa9c2" exitCode=0 Dec 12 08:27:59 crc kubenswrapper[4867]: I1212 08:27:59.707022 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerDied","Data":"d1d728f65936f38fd0eb61a1d97ab7f7117070ec9ef659662b4bb39409baa9c2"} Dec 12 08:27:59 crc kubenswrapper[4867]: I1212 08:27:59.721276 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d5d9f84d-gcsp7" podUID="a52ae013-bec4-4e49-8322-9f291818039d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.99:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 12 08:28:00 crc kubenswrapper[4867]: I1212 08:28:00.718456 4867 generic.go:334] "Generic (PLEG): container finished" podID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerID="02ca73c39aa514d6356733a645283aac939ef1d501c9d35fbb370778c5fb8509" exitCode=0 Dec 12 08:28:00 crc kubenswrapper[4867]: I1212 08:28:00.718498 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerDied","Data":"02ca73c39aa514d6356733a645283aac939ef1d501c9d35fbb370778c5fb8509"} Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.367211 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.367557 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h65wz" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="registry-server" containerID="cri-o://f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006" gracePeriod=2 Dec 12 08:28:01 crc kubenswrapper[4867]: E1212 08:28:01.599691 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c021602_59a6_4abe_bb01_e2ccc2a0926d.slice/crio-f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c021602_59a6_4abe_bb01_e2ccc2a0926d.slice/crio-conmon-f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.729689 4867 generic.go:334] "Generic (PLEG): container finished" podID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerID="f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006" exitCode=0 Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.729879 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerDied","Data":"f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006"} Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.815537 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.928330 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content\") pod \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.928453 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mkxt\" (UniqueName: \"kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt\") pod \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.928526 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities\") pod \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\" (UID: \"0c021602-59a6-4abe-bb01-e2ccc2a0926d\") " Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.931338 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities" (OuterVolumeSpecName: "utilities") pod "0c021602-59a6-4abe-bb01-e2ccc2a0926d" (UID: "0c021602-59a6-4abe-bb01-e2ccc2a0926d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:28:01 crc kubenswrapper[4867]: I1212 08:28:01.940035 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt" (OuterVolumeSpecName: "kube-api-access-6mkxt") pod "0c021602-59a6-4abe-bb01-e2ccc2a0926d" (UID: "0c021602-59a6-4abe-bb01-e2ccc2a0926d"). InnerVolumeSpecName "kube-api-access-6mkxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.030663 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.030976 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mkxt\" (UniqueName: \"kubernetes.io/projected/0c021602-59a6-4abe-bb01-e2ccc2a0926d-kube-api-access-6mkxt\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.046814 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.057688 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c021602-59a6-4abe-bb01-e2ccc2a0926d" (UID: "0c021602-59a6-4abe-bb01-e2ccc2a0926d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.132871 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle\") pod \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.133081 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q58sw\" (UniqueName: \"kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw\") pod \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.133349 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util\") pod \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\" (UID: \"a27475d4-d4ab-4f25-b83c-d107efa67cd7\") " Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.133956 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c021602-59a6-4abe-bb01-e2ccc2a0926d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.137438 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle" (OuterVolumeSpecName: "bundle") pod "a27475d4-d4ab-4f25-b83c-d107efa67cd7" (UID: "a27475d4-d4ab-4f25-b83c-d107efa67cd7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.137465 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw" (OuterVolumeSpecName: "kube-api-access-q58sw") pod "a27475d4-d4ab-4f25-b83c-d107efa67cd7" (UID: "a27475d4-d4ab-4f25-b83c-d107efa67cd7"). InnerVolumeSpecName "kube-api-access-q58sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.147293 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util" (OuterVolumeSpecName: "util") pod "a27475d4-d4ab-4f25-b83c-d107efa67cd7" (UID: "a27475d4-d4ab-4f25-b83c-d107efa67cd7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.235445 4867 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.235499 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q58sw\" (UniqueName: \"kubernetes.io/projected/a27475d4-d4ab-4f25-b83c-d107efa67cd7-kube-api-access-q58sw\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.235515 4867 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a27475d4-d4ab-4f25-b83c-d107efa67cd7-util\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.739429 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" event={"ID":"a27475d4-d4ab-4f25-b83c-d107efa67cd7","Type":"ContainerDied","Data":"83f1aaa0e2d88cd3de9758ecb82af2baeba1bbfa440a6466dd2cfdca79a60c15"} Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.739478 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f1aaa0e2d88cd3de9758ecb82af2baeba1bbfa440a6466dd2cfdca79a60c15" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.739516 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.741999 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h65wz" event={"ID":"0c021602-59a6-4abe-bb01-e2ccc2a0926d","Type":"ContainerDied","Data":"194ee2518bc08a697f64fc1dab74460e96ec61fb71f91283304029d6fa9edd97"} Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.742050 4867 scope.go:117] "RemoveContainer" containerID="f4aa8a55e860ee37b65717430361d6862916aa428ca82aeec650b350db646006" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.742058 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h65wz" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.777898 4867 scope.go:117] "RemoveContainer" containerID="b890f658fc9de0b0ff0feedd3b93577c57000156fed9605aefb5775f7b886337" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.780420 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.793595 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h65wz"] Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.798829 4867 scope.go:117] "RemoveContainer" containerID="ac3f2ddd560a04048992597f7203994f6cbf5fa3f6d152da5b883bb8e9c25e52" Dec 12 08:28:02 crc kubenswrapper[4867]: I1212 08:28:02.850301 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" path="/var/lib/kubelet/pods/0c021602-59a6-4abe-bb01-e2ccc2a0926d/volumes" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.427094 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh"] Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428165 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="util" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428178 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="util" Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428193 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="extract-content" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428199 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="extract-content" Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428215 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="registry-server" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428239 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="registry-server" Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428249 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="pull" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428255 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="pull" Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428275 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="extract-utilities" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428281 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="extract-utilities" Dec 12 08:28:11 crc kubenswrapper[4867]: E1212 08:28:11.428300 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="extract" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428306 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="extract" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428481 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a27475d4-d4ab-4f25-b83c-d107efa67cd7" containerName="extract" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.428498 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c021602-59a6-4abe-bb01-e2ccc2a0926d" containerName="registry-server" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.429160 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.431197 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.431498 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.447873 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.448910 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-dzrt5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.494727 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.496448 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.499077 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.499274 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-csns5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.513174 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.519741 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd2kk\" (UniqueName: \"kubernetes.io/projected/751ded8c-1d36-4275-a0b5-d6444285b50b-kube-api-access-gd2kk\") pod \"obo-prometheus-operator-668cf9dfbb-df5xh\" (UID: \"751ded8c-1d36-4275-a0b5-d6444285b50b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.528736 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.530249 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.542807 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.622940 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.623055 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.623108 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.623193 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.623285 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd2kk\" (UniqueName: \"kubernetes.io/projected/751ded8c-1d36-4275-a0b5-d6444285b50b-kube-api-access-gd2kk\") pod \"obo-prometheus-operator-668cf9dfbb-df5xh\" (UID: \"751ded8c-1d36-4275-a0b5-d6444285b50b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.647509 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd2kk\" (UniqueName: \"kubernetes.io/projected/751ded8c-1d36-4275-a0b5-d6444285b50b-kube-api-access-gd2kk\") pod \"obo-prometheus-operator-668cf9dfbb-df5xh\" (UID: \"751ded8c-1d36-4275-a0b5-d6444285b50b\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.669589 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-56mq5"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.671003 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.674355 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-tcwdt" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.674722 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.687381 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-56mq5"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.725988 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.726081 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.726123 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.726193 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.726255 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/035114f3-82c3-4566-83ab-7b4151b7232d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.726303 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8nnz\" (UniqueName: \"kubernetes.io/projected/035114f3-82c3-4566-83ab-7b4151b7232d-kube-api-access-x8nnz\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.730703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.730886 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.737702 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3bc2cd37-97e9-43b8-a499-0b706dcc7fd6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-rfr5n\" (UID: \"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.750407 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.752977 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b9258b12-9a2e-4219-aaa4-1ec3717692b2-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-98568f549-k4lz7\" (UID: \"b9258b12-9a2e-4219-aaa4-1ec3717692b2\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.816484 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.837619 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/035114f3-82c3-4566-83ab-7b4151b7232d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.837722 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8nnz\" (UniqueName: \"kubernetes.io/projected/035114f3-82c3-4566-83ab-7b4151b7232d-kube-api-access-x8nnz\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.862286 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/035114f3-82c3-4566-83ab-7b4151b7232d-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.862299 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.870036 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8nnz\" (UniqueName: \"kubernetes.io/projected/035114f3-82c3-4566-83ab-7b4151b7232d-kube-api-access-x8nnz\") pod \"observability-operator-d8bb48f5d-56mq5\" (UID: \"035114f3-82c3-4566-83ab-7b4151b7232d\") " pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.912723 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-kgxfk"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.914212 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.921149 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4rdbg" Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.938192 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-kgxfk"] Dec 12 08:28:11 crc kubenswrapper[4867]: I1212 08:28:11.976494 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.046964 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl62g\" (UniqueName: \"kubernetes.io/projected/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-kube-api-access-cl62g\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.047832 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-openshift-service-ca\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.152788 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-openshift-service-ca\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.152870 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl62g\" (UniqueName: \"kubernetes.io/projected/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-kube-api-access-cl62g\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.153895 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-openshift-service-ca\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.178454 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl62g\" (UniqueName: \"kubernetes.io/projected/545d1312-b5d7-40fa-85c4-d54c8e09c7a2-kube-api-access-cl62g\") pod \"perses-operator-5446b9c989-kgxfk\" (UID: \"545d1312-b5d7-40fa-85c4-d54c8e09c7a2\") " pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.303581 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.366369 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh"] Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.579021 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n"] Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.660653 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7"] Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.814161 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-56mq5"] Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.877644 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" event={"ID":"035114f3-82c3-4566-83ab-7b4151b7232d","Type":"ContainerStarted","Data":"93b50e6e2330a28baf287e4675ea9bc9566a948b3aba37f943d497555294ae12"} Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.880032 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" event={"ID":"b9258b12-9a2e-4219-aaa4-1ec3717692b2","Type":"ContainerStarted","Data":"00198568c2f11acb33889798b6beb3b982500861cbc85969be4d1b6313d6b3ae"} Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.881118 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" event={"ID":"751ded8c-1d36-4275-a0b5-d6444285b50b","Type":"ContainerStarted","Data":"29d7727d255eb61dbe69febfacc9b6853c311bfbb84cfe49e97d46033ca3fed2"} Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.882393 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" event={"ID":"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6","Type":"ContainerStarted","Data":"4e9affce4e97d3c0f2734048c072117da3714ee33ff124b5a6346613d48da80b"} Dec 12 08:28:12 crc kubenswrapper[4867]: I1212 08:28:12.920880 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-kgxfk"] Dec 12 08:28:12 crc kubenswrapper[4867]: W1212 08:28:12.923497 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod545d1312_b5d7_40fa_85c4_d54c8e09c7a2.slice/crio-8aba4d8f4bdbde8a0773f900de2e8e503ca93e18bed6455872ac326fec9e05cd WatchSource:0}: Error finding container 8aba4d8f4bdbde8a0773f900de2e8e503ca93e18bed6455872ac326fec9e05cd: Status 404 returned error can't find the container with id 8aba4d8f4bdbde8a0773f900de2e8e503ca93e18bed6455872ac326fec9e05cd Dec 12 08:28:13 crc kubenswrapper[4867]: I1212 08:28:13.894942 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" event={"ID":"545d1312-b5d7-40fa-85c4-d54c8e09c7a2","Type":"ContainerStarted","Data":"8aba4d8f4bdbde8a0773f900de2e8e503ca93e18bed6455872ac326fec9e05cd"} Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.043325 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-27d8l"] Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.061325 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1932-account-create-update-kgphl"] Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.070866 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-27d8l"] Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.084734 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1932-account-create-update-kgphl"] Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.849867 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e3e4886-12ff-4e96-b20e-7c13c2d54ce1" path="/var/lib/kubelet/pods/0e3e4886-12ff-4e96-b20e-7c13c2d54ce1/volumes" Dec 12 08:28:16 crc kubenswrapper[4867]: I1212 08:28:16.852241 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7e6e15-71a3-41f0-a41e-4eae2d0fec66" path="/var/lib/kubelet/pods/ed7e6e15-71a3-41f0-a41e-4eae2d0fec66/volumes" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.027925 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" event={"ID":"035114f3-82c3-4566-83ab-7b4151b7232d","Type":"ContainerStarted","Data":"33f3a6165d0043297fbf71bac8dd4bc59ebcf512f9dc29571021f1652c4729aa"} Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.028556 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.040791 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" event={"ID":"b9258b12-9a2e-4219-aaa4-1ec3717692b2","Type":"ContainerStarted","Data":"f23024ca6a4a99bca253cc2b85b1f3f2b42d07962b081aee94c4fc44c3b3bc53"} Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.050974 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" event={"ID":"751ded8c-1d36-4275-a0b5-d6444285b50b","Type":"ContainerStarted","Data":"3ee54466b64a4edeaa9b9f4890a8bdb385fc7ebf38e1e11501ad55d4e5704d46"} Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.053747 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" event={"ID":"3bc2cd37-97e9-43b8-a499-0b706dcc7fd6","Type":"ContainerStarted","Data":"c28057b3c30d1eb75c9bd75f801464763047610c49165f647f07a793a7c9f955"} Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.058060 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" podStartSLOduration=2.738690271 podStartE2EDuration="11.058043975s" podCreationTimestamp="2025-12-12 08:28:11 +0000 UTC" firstStartedPulling="2025-12-12 08:28:12.800996474 +0000 UTC m=+5980.372377743" lastFinishedPulling="2025-12-12 08:28:21.120350178 +0000 UTC m=+5988.691731447" observedRunningTime="2025-12-12 08:28:22.057966543 +0000 UTC m=+5989.629347832" watchObservedRunningTime="2025-12-12 08:28:22.058043975 +0000 UTC m=+5989.629425244" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.061218 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" event={"ID":"545d1312-b5d7-40fa-85c4-d54c8e09c7a2","Type":"ContainerStarted","Data":"beda5fb43d063d3379517930c5bdf43d5cf58a884010190e52f87ad7daf62e3a"} Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.061502 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.081847 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-df5xh" podStartSLOduration=2.508582663 podStartE2EDuration="11.081822204s" podCreationTimestamp="2025-12-12 08:28:11 +0000 UTC" firstStartedPulling="2025-12-12 08:28:12.403964696 +0000 UTC m=+5979.975345965" lastFinishedPulling="2025-12-12 08:28:20.977204237 +0000 UTC m=+5988.548585506" observedRunningTime="2025-12-12 08:28:22.07438397 +0000 UTC m=+5989.645765239" watchObservedRunningTime="2025-12-12 08:28:22.081822204 +0000 UTC m=+5989.653203473" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.118066 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-56mq5" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.118960 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-rfr5n" podStartSLOduration=2.741971513 podStartE2EDuration="11.118947126s" podCreationTimestamp="2025-12-12 08:28:11 +0000 UTC" firstStartedPulling="2025-12-12 08:28:12.598412229 +0000 UTC m=+5980.169793498" lastFinishedPulling="2025-12-12 08:28:20.975387842 +0000 UTC m=+5988.546769111" observedRunningTime="2025-12-12 08:28:22.097655528 +0000 UTC m=+5989.669036797" watchObservedRunningTime="2025-12-12 08:28:22.118947126 +0000 UTC m=+5989.690328395" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.187574 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-98568f549-k4lz7" podStartSLOduration=2.839294796 podStartE2EDuration="11.187552397s" podCreationTimestamp="2025-12-12 08:28:11 +0000 UTC" firstStartedPulling="2025-12-12 08:28:12.628496445 +0000 UTC m=+5980.199877724" lastFinishedPulling="2025-12-12 08:28:20.976754056 +0000 UTC m=+5988.548135325" observedRunningTime="2025-12-12 08:28:22.168377611 +0000 UTC m=+5989.739758880" watchObservedRunningTime="2025-12-12 08:28:22.187552397 +0000 UTC m=+5989.758933666" Dec 12 08:28:22 crc kubenswrapper[4867]: I1212 08:28:22.200870 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" podStartSLOduration=3.152454334 podStartE2EDuration="11.200824336s" podCreationTimestamp="2025-12-12 08:28:11 +0000 UTC" firstStartedPulling="2025-12-12 08:28:12.926745653 +0000 UTC m=+5980.498126922" lastFinishedPulling="2025-12-12 08:28:20.975115655 +0000 UTC m=+5988.546496924" observedRunningTime="2025-12-12 08:28:22.193287459 +0000 UTC m=+5989.764668728" watchObservedRunningTime="2025-12-12 08:28:22.200824336 +0000 UTC m=+5989.772205605" Dec 12 08:28:28 crc kubenswrapper[4867]: I1212 08:28:28.988492 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:28:28 crc kubenswrapper[4867]: I1212 08:28:28.988824 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:28:28 crc kubenswrapper[4867]: I1212 08:28:28.988859 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:28:28 crc kubenswrapper[4867]: I1212 08:28:28.989654 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:28:28 crc kubenswrapper[4867]: I1212 08:28:28.989713 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" gracePeriod=600 Dec 12 08:28:30 crc kubenswrapper[4867]: I1212 08:28:30.132531 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" exitCode=0 Dec 12 08:28:30 crc kubenswrapper[4867]: I1212 08:28:30.132638 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098"} Dec 12 08:28:30 crc kubenswrapper[4867]: I1212 08:28:30.132908 4867 scope.go:117] "RemoveContainer" containerID="efde264d54ef8da69fa26a189e794083566dff8e5916bcbc69d18b673355bd52" Dec 12 08:28:30 crc kubenswrapper[4867]: E1212 08:28:30.422884 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:28:31 crc kubenswrapper[4867]: I1212 08:28:31.145566 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:28:31 crc kubenswrapper[4867]: E1212 08:28:31.146180 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:28:32 crc kubenswrapper[4867]: I1212 08:28:32.307798 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.266938 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.267717 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" containerName="openstackclient" containerID="cri-o://0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd" gracePeriod=2 Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.301782 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.318170 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 12 08:28:35 crc kubenswrapper[4867]: E1212 08:28:35.318734 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" containerName="openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.318758 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" containerName="openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.319002 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" containerName="openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.319924 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.328456 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.351176 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.396961 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.397269 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.397426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd5fc\" (UniqueName: \"kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.397496 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.451361 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.453019 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.456563 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-n2vjt" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.480457 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.499506 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.499635 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.499678 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd5fc\" (UniqueName: \"kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.499703 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.508410 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.537087 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.539649 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd5fc\" (UniqueName: \"kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.545732 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret\") pod \"openstackclient\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.602950 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkssn\" (UniqueName: \"kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn\") pod \"kube-state-metrics-0\" (UID: \"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d\") " pod="openstack/kube-state-metrics-0" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.653844 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.709887 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkssn\" (UniqueName: \"kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn\") pod \"kube-state-metrics-0\" (UID: \"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d\") " pod="openstack/kube-state-metrics-0" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.738111 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkssn\" (UniqueName: \"kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn\") pod \"kube-state-metrics-0\" (UID: \"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d\") " pod="openstack/kube-state-metrics-0" Dec 12 08:28:35 crc kubenswrapper[4867]: I1212 08:28:35.781897 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.222954 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.240138 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.241026 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.252386 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.252461 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.252695 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-fs6cj" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.252856 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.252991 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337426 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337488 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjcjv\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-kube-api-access-sjcjv\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337531 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337573 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337617 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337642 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.337794 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440495 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440590 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440619 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjcjv\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-kube-api-access-sjcjv\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440655 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440697 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440741 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.440766 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.441767 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.446593 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.446604 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.449346 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.449840 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/b95f4529-9df6-48cf-a611-4b0b049ca8f4-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.450889 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.515289 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjcjv\" (UniqueName: \"kubernetes.io/projected/b95f4529-9df6-48cf-a611-4b0b049ca8f4-kube-api-access-sjcjv\") pod \"alertmanager-metric-storage-0\" (UID: \"b95f4529-9df6-48cf-a611-4b0b049ca8f4\") " pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.563316 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:28:36 crc kubenswrapper[4867]: W1212 08:28:36.563461 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d97ed7f_b8eb_4a0e_983d_71d9aa20648d.slice/crio-81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb WatchSource:0}: Error finding container 81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb: Status 404 returned error can't find the container with id 81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.601051 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.668904 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.814601 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.819618 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.823752 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.824157 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.824354 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.824508 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.825367 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.825518 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jdw6f" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.877069 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970675 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970714 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970784 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970902 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmr2\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970961 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.970985 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.971032 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:36 crc kubenswrapper[4867]: I1212 08:28:36.971586 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073273 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073325 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073381 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmr2\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073555 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073578 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073614 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.073668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.075462 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.091095 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.091151 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c9dd10525c976faca9fd325689dfb274363330ecf89d10faba3f29d2a60e6436/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.091705 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.093691 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.093769 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.095051 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.098772 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmr2\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.101265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.166016 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.285580 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"42806f5a-b1ec-47a6-b55e-494f7bfa8108","Type":"ContainerStarted","Data":"ed1850cfd960e8db4ec02f08d50e9b23ce0f12d99a0428a1025b86a9a22517d8"} Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.285640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"42806f5a-b1ec-47a6-b55e-494f7bfa8108","Type":"ContainerStarted","Data":"734b6b9f8f7eaf7667caacdde844d352919a9f8d6c0f0f204b99ead1ba7d0940"} Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.299247 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d","Type":"ContainerStarted","Data":"81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb"} Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.317746 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.317726 podStartE2EDuration="2.317726s" podCreationTimestamp="2025-12-12 08:28:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:28:37.316354986 +0000 UTC m=+6004.887736275" watchObservedRunningTime="2025-12-12 08:28:37.317726 +0000 UTC m=+6004.889107269" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.347894 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 12 08:28:37 crc kubenswrapper[4867]: W1212 08:28:37.358136 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb95f4529_9df6_48cf_a611_4b0b049ca8f4.slice/crio-67dcf87193974fa0c02e27340662783d802dfb57a7f0feebb134e5a230b25622 WatchSource:0}: Error finding container 67dcf87193974fa0c02e27340662783d802dfb57a7f0feebb134e5a230b25622: Status 404 returned error can't find the container with id 67dcf87193974fa0c02e27340662783d802dfb57a7f0feebb134e5a230b25622 Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.463447 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.761399 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.804153 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config\") pod \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.804252 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l252f\" (UniqueName: \"kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f\") pod \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.804326 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle\") pod \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.804498 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret\") pod \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\" (UID: \"07ec8a92-18c0-4dae-9f03-775fe99bcb01\") " Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.815624 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f" (OuterVolumeSpecName: "kube-api-access-l252f") pod "07ec8a92-18c0-4dae-9f03-775fe99bcb01" (UID: "07ec8a92-18c0-4dae-9f03-775fe99bcb01"). InnerVolumeSpecName "kube-api-access-l252f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.873020 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "07ec8a92-18c0-4dae-9f03-775fe99bcb01" (UID: "07ec8a92-18c0-4dae-9f03-775fe99bcb01"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.885773 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "07ec8a92-18c0-4dae-9f03-775fe99bcb01" (UID: "07ec8a92-18c0-4dae-9f03-775fe99bcb01"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.890950 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07ec8a92-18c0-4dae-9f03-775fe99bcb01" (UID: "07ec8a92-18c0-4dae-9f03-775fe99bcb01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.906571 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.906604 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l252f\" (UniqueName: \"kubernetes.io/projected/07ec8a92-18c0-4dae-9f03-775fe99bcb01-kube-api-access-l252f\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.906616 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:37 crc kubenswrapper[4867]: I1212 08:28:37.906626 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07ec8a92-18c0-4dae-9f03-775fe99bcb01-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.137889 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.310346 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d","Type":"ContainerStarted","Data":"dcf41a6869ed2e886079b80edbb9d4a3f839e14309b6092fa8811d6f95c5e8c0"} Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.310914 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.312214 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerStarted","Data":"d3489583922108b0b5f1fb526c1dbe6953c38ebe8a0fa7b877233d49798f91c6"} Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.314378 4867 generic.go:334] "Generic (PLEG): container finished" podID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" containerID="0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd" exitCode=137 Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.314451 4867 scope.go:117] "RemoveContainer" containerID="0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.314492 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.316462 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b95f4529-9df6-48cf-a611-4b0b049ca8f4","Type":"ContainerStarted","Data":"67dcf87193974fa0c02e27340662783d802dfb57a7f0feebb134e5a230b25622"} Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.332368 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.754357512 podStartE2EDuration="3.332352988s" podCreationTimestamp="2025-12-12 08:28:35 +0000 UTC" firstStartedPulling="2025-12-12 08:28:36.58172157 +0000 UTC m=+6004.153102849" lastFinishedPulling="2025-12-12 08:28:37.159717056 +0000 UTC m=+6004.731098325" observedRunningTime="2025-12-12 08:28:38.328870532 +0000 UTC m=+6005.900251801" watchObservedRunningTime="2025-12-12 08:28:38.332352988 +0000 UTC m=+6005.903734257" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.337876 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.353385 4867 scope.go:117] "RemoveContainer" containerID="0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd" Dec 12 08:28:38 crc kubenswrapper[4867]: E1212 08:28:38.353904 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd\": container with ID starting with 0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd not found: ID does not exist" containerID="0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.353940 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd"} err="failed to get container status \"0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd\": rpc error: code = NotFound desc = could not find container \"0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd\": container with ID starting with 0d8d84ef047cc05ff8e478c5e2a548b8fbccdedc6caf5f22a41e01bc2bd513dd not found: ID does not exist" Dec 12 08:28:38 crc kubenswrapper[4867]: I1212 08:28:38.855686 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ec8a92-18c0-4dae-9f03-775fe99bcb01" path="/var/lib/kubelet/pods/07ec8a92-18c0-4dae-9f03-775fe99bcb01/volumes" Dec 12 08:28:41 crc kubenswrapper[4867]: I1212 08:28:41.838518 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:28:41 crc kubenswrapper[4867]: E1212 08:28:41.839142 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:28:44 crc kubenswrapper[4867]: I1212 08:28:44.409813 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b95f4529-9df6-48cf-a611-4b0b049ca8f4","Type":"ContainerStarted","Data":"efcd5be5eb3d84671bb8b1c833f64b34d5b12bf6db03e24c643ca146429258d5"} Dec 12 08:28:44 crc kubenswrapper[4867]: I1212 08:28:44.411962 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerStarted","Data":"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf"} Dec 12 08:28:45 crc kubenswrapper[4867]: I1212 08:28:45.788778 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 12 08:28:51 crc kubenswrapper[4867]: I1212 08:28:51.476998 4867 generic.go:334] "Generic (PLEG): container finished" podID="b95f4529-9df6-48cf-a611-4b0b049ca8f4" containerID="efcd5be5eb3d84671bb8b1c833f64b34d5b12bf6db03e24c643ca146429258d5" exitCode=0 Dec 12 08:28:51 crc kubenswrapper[4867]: I1212 08:28:51.477125 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b95f4529-9df6-48cf-a611-4b0b049ca8f4","Type":"ContainerDied","Data":"efcd5be5eb3d84671bb8b1c833f64b34d5b12bf6db03e24c643ca146429258d5"} Dec 12 08:28:51 crc kubenswrapper[4867]: I1212 08:28:51.479304 4867 generic.go:334] "Generic (PLEG): container finished" podID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerID="d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf" exitCode=0 Dec 12 08:28:51 crc kubenswrapper[4867]: I1212 08:28:51.479359 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerDied","Data":"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf"} Dec 12 08:28:53 crc kubenswrapper[4867]: I1212 08:28:53.837925 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:28:53 crc kubenswrapper[4867]: E1212 08:28:53.838777 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:28:55 crc kubenswrapper[4867]: I1212 08:28:55.322326 4867 scope.go:117] "RemoveContainer" containerID="4d2fb0d826d4a9896c79df299757c9549de17bdd402f90896ffc21c6a125bcc2" Dec 12 08:28:59 crc kubenswrapper[4867]: I1212 08:28:59.045034 4867 scope.go:117] "RemoveContainer" containerID="791d9cdf6c3dfa77fa27cfa1d58814de36255d73326ecaafc877cf3bba8c3e96" Dec 12 08:28:59 crc kubenswrapper[4867]: I1212 08:28:59.570862 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerStarted","Data":"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6"} Dec 12 08:28:59 crc kubenswrapper[4867]: I1212 08:28:59.575805 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b95f4529-9df6-48cf-a611-4b0b049ca8f4","Type":"ContainerStarted","Data":"53f928b9853a43ce127b687d785b132edaf2aa21faf89861066f79f057b2abff"} Dec 12 08:29:03 crc kubenswrapper[4867]: I1212 08:29:03.614824 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"b95f4529-9df6-48cf-a611-4b0b049ca8f4","Type":"ContainerStarted","Data":"ff8bb4a7365acb39cd3d9b23cf2081d0c633efade7a1b0f9d3443fad628cc986"} Dec 12 08:29:03 crc kubenswrapper[4867]: I1212 08:29:03.615816 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 12 08:29:03 crc kubenswrapper[4867]: I1212 08:29:03.618127 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 12 08:29:03 crc kubenswrapper[4867]: I1212 08:29:03.648859 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.90528373 podStartE2EDuration="27.648835847s" podCreationTimestamp="2025-12-12 08:28:36 +0000 UTC" firstStartedPulling="2025-12-12 08:28:37.387396043 +0000 UTC m=+6004.958777312" lastFinishedPulling="2025-12-12 08:28:59.13094816 +0000 UTC m=+6026.702329429" observedRunningTime="2025-12-12 08:29:03.638066029 +0000 UTC m=+6031.209447298" watchObservedRunningTime="2025-12-12 08:29:03.648835847 +0000 UTC m=+6031.220217116" Dec 12 08:29:04 crc kubenswrapper[4867]: I1212 08:29:04.628432 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerStarted","Data":"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c"} Dec 12 08:29:07 crc kubenswrapper[4867]: I1212 08:29:07.663947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerStarted","Data":"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236"} Dec 12 08:29:07 crc kubenswrapper[4867]: I1212 08:29:07.700936 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.006071586 podStartE2EDuration="32.70091304s" podCreationTimestamp="2025-12-12 08:28:35 +0000 UTC" firstStartedPulling="2025-12-12 08:28:38.151589885 +0000 UTC m=+6005.722971144" lastFinishedPulling="2025-12-12 08:29:06.846431329 +0000 UTC m=+6034.417812598" observedRunningTime="2025-12-12 08:29:07.691321602 +0000 UTC m=+6035.262702871" watchObservedRunningTime="2025-12-12 08:29:07.70091304 +0000 UTC m=+6035.272294309" Dec 12 08:29:07 crc kubenswrapper[4867]: I1212 08:29:07.838262 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:29:07 crc kubenswrapper[4867]: E1212 08:29:07.838862 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:29:10 crc kubenswrapper[4867]: I1212 08:29:10.044340 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vpzt4"] Dec 12 08:29:10 crc kubenswrapper[4867]: I1212 08:29:10.053975 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vpzt4"] Dec 12 08:29:10 crc kubenswrapper[4867]: I1212 08:29:10.852622 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f2c606-4920-493b-a23d-b9b5209f3636" path="/var/lib/kubelet/pods/17f2c606-4920-493b-a23d-b9b5209f3636/volumes" Dec 12 08:29:12 crc kubenswrapper[4867]: I1212 08:29:12.464845 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.709507 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.712847 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.714819 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.715528 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.731836 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.857997 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858053 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858076 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858097 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858124 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858141 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hbb4\" (UniqueName: \"kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.858202 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.959963 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960020 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960048 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960072 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960106 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960123 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hbb4\" (UniqueName: \"kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.960174 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.961327 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.961346 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.969946 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.969966 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.970315 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.970612 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:17 crc kubenswrapper[4867]: I1212 08:29:17.981193 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hbb4\" (UniqueName: \"kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4\") pod \"ceilometer-0\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " pod="openstack/ceilometer-0" Dec 12 08:29:18 crc kubenswrapper[4867]: I1212 08:29:18.034388 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:29:18 crc kubenswrapper[4867]: I1212 08:29:18.679409 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:18 crc kubenswrapper[4867]: I1212 08:29:18.785133 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerStarted","Data":"19ed42d4d724614fe348df2b0c45dbd9a9b66414095b8903b923f94876560f24"} Dec 12 08:29:18 crc kubenswrapper[4867]: I1212 08:29:18.837985 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:29:18 crc kubenswrapper[4867]: E1212 08:29:18.838511 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:29:22 crc kubenswrapper[4867]: I1212 08:29:22.464993 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:22 crc kubenswrapper[4867]: I1212 08:29:22.467545 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:22 crc kubenswrapper[4867]: I1212 08:29:22.860161 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.184963 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.186056 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" containerName="openstackclient" containerID="cri-o://ed1850cfd960e8db4ec02f08d50e9b23ce0f12d99a0428a1025b86a9a22517d8" gracePeriod=2 Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.205716 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.217718 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 12 08:29:24 crc kubenswrapper[4867]: E1212 08:29:24.218574 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" containerName="openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.218687 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" containerName="openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.219047 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" containerName="openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.220446 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.236209 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.241826 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" podUID="07f2128b-4e78-430a-9af7-45cf3e402573" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.302695 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vw2\" (UniqueName: \"kubernetes.io/projected/07f2128b-4e78-430a-9af7-45cf3e402573-kube-api-access-c2vw2\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.302786 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config-secret\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.302846 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.302977 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.405601 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.405895 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vw2\" (UniqueName: \"kubernetes.io/projected/07f2128b-4e78-430a-9af7-45cf3e402573-kube-api-access-c2vw2\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.405977 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config-secret\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.406044 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.407869 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.412274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-combined-ca-bundle\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.413113 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/07f2128b-4e78-430a-9af7-45cf3e402573-openstack-config-secret\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.433807 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vw2\" (UniqueName: \"kubernetes.io/projected/07f2128b-4e78-430a-9af7-45cf3e402573-kube-api-access-c2vw2\") pod \"openstackclient\" (UID: \"07f2128b-4e78-430a-9af7-45cf3e402573\") " pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.555668 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:29:24 crc kubenswrapper[4867]: I1212 08:29:24.874433 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerStarted","Data":"303fec35adcf223befd5fb8ce60bcf535b8bda87ed35edd1d2da82fc96a72db6"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.250238 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.452529 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.453196 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="prometheus" containerID="cri-o://d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6" gracePeriod=600 Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.453373 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="thanos-sidecar" containerID="cri-o://c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236" gracePeriod=600 Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.453397 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="config-reloader" containerID="cri-o://2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c" gracePeriod=600 Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.883921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerStarted","Data":"c1d30a17c8ae74d8ff1b6da24997ec31f28503af3d802503cd87a92421f710eb"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.886121 4867 generic.go:334] "Generic (PLEG): container finished" podID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerID="c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236" exitCode=0 Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.886151 4867 generic.go:334] "Generic (PLEG): container finished" podID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerID="d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6" exitCode=0 Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.886171 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerDied","Data":"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.886201 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerDied","Data":"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.887268 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"07f2128b-4e78-430a-9af7-45cf3e402573","Type":"ContainerStarted","Data":"534d944848c6db5c8dc19c5e641e09cc596eb45fc4b401109296008da42f9c7b"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.887292 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"07f2128b-4e78-430a-9af7-45cf3e402573","Type":"ContainerStarted","Data":"17d49a4f62d89734cad22754d9fb7805a8b35d3983547c7ebc2053083a631d21"} Dec 12 08:29:25 crc kubenswrapper[4867]: I1212 08:29:25.902453 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.902431668 podStartE2EDuration="1.902431668s" podCreationTimestamp="2025-12-12 08:29:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:29:25.899167647 +0000 UTC m=+6053.470548916" watchObservedRunningTime="2025-12-12 08:29:25.902431668 +0000 UTC m=+6053.473812947" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.760831 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.852845 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.852958 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853064 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853144 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853176 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853257 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vmr2\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.853377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"815df45e-d1b3-4d94-9054-0ccc8f427af4\" (UID: \"815df45e-d1b3-4d94-9054-0ccc8f427af4\") " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.854025 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.860463 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out" (OuterVolumeSpecName: "config-out") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.861406 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config" (OuterVolumeSpecName: "config") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.861614 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2" (OuterVolumeSpecName: "kube-api-access-4vmr2") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "kube-api-access-4vmr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.863317 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.864182 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.903690 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.903903 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config" (OuterVolumeSpecName: "web-config") pod "815df45e-d1b3-4d94-9054-0ccc8f427af4" (UID: "815df45e-d1b3-4d94-9054-0ccc8f427af4"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.935132 4867 generic.go:334] "Generic (PLEG): container finished" podID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerID="2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c" exitCode=0 Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.935379 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.943905 4867 generic.go:334] "Generic (PLEG): container finished" podID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" containerID="ed1850cfd960e8db4ec02f08d50e9b23ce0f12d99a0428a1025b86a9a22517d8" exitCode=137 Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955788 4867 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955834 4867 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/815df45e-d1b3-4d94-9054-0ccc8f427af4-config-out\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955847 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955859 4867 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/815df45e-d1b3-4d94-9054-0ccc8f427af4-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955874 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vmr2\" (UniqueName: \"kubernetes.io/projected/815df45e-d1b3-4d94-9054-0ccc8f427af4-kube-api-access-4vmr2\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955889 4867 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-web-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955923 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") on node \"crc\" " Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.955938 4867 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/815df45e-d1b3-4d94-9054-0ccc8f427af4-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.958548 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerDied","Data":"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c"} Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.958592 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"815df45e-d1b3-4d94-9054-0ccc8f427af4","Type":"ContainerDied","Data":"d3489583922108b0b5f1fb526c1dbe6953c38ebe8a0fa7b877233d49798f91c6"} Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.958613 4867 scope.go:117] "RemoveContainer" containerID="c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236" Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.988652 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:26 crc kubenswrapper[4867]: I1212 08:29:26.997666 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.007716 4867 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.007885 4867 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b") on node "crc" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032097 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.032586 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="thanos-sidecar" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032604 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="thanos-sidecar" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.032648 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="prometheus" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032657 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="prometheus" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.032667 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="init-config-reloader" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032673 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="init-config-reloader" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.032685 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="config-reloader" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032691 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="config-reloader" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032907 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="thanos-sidecar" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032922 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="prometheus" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.032946 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" containerName="config-reloader" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.035611 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050319 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050486 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050561 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050669 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jdw6f" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050749 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.050760 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.057981 4867 reconciler_common.go:293] "Volume detached for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.066600 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.069440 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.070179 4867 scope.go:117] "RemoveContainer" containerID="2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.159500 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.159904 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.159983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160005 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6vwq\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-kube-api-access-l6vwq\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160024 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160040 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160079 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160126 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160147 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160165 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.160207 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.188497 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.196429 4867 scope.go:117] "RemoveContainer" containerID="d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261400 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261473 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261497 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6vwq\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-kube-api-access-l6vwq\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261516 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261536 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261579 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261627 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261653 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261673 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261703 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.261748 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.266403 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.267763 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.268917 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.270352 4867 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.270382 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c9dd10525c976faca9fd325689dfb274363330ecf89d10faba3f29d2a60e6436/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.282404 4867 scope.go:117] "RemoveContainer" containerID="d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.282860 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.283133 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.293313 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6vwq\" (UniqueName: \"kubernetes.io/projected/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-kube-api-access-l6vwq\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.295042 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.295872 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.296414 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.296808 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/6d045ee8-6266-4f6f-be38-0a8d39d1fb7e-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.364979 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd5fc\" (UniqueName: \"kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc\") pod \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.365045 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config\") pod \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.365075 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret\") pod \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.365150 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle\") pod \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\" (UID: \"42806f5a-b1ec-47a6-b55e-494f7bfa8108\") " Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.377382 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc" (OuterVolumeSpecName: "kube-api-access-jd5fc") pod "42806f5a-b1ec-47a6-b55e-494f7bfa8108" (UID: "42806f5a-b1ec-47a6-b55e-494f7bfa8108"). InnerVolumeSpecName "kube-api-access-jd5fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.378538 4867 scope.go:117] "RemoveContainer" containerID="c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.381100 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236\": container with ID starting with c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236 not found: ID does not exist" containerID="c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.381140 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236"} err="failed to get container status \"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236\": rpc error: code = NotFound desc = could not find container \"c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236\": container with ID starting with c9e73b157c1b2c3375057388cbe75647b753f352bbeef1572a68a5fde606f236 not found: ID does not exist" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.381161 4867 scope.go:117] "RemoveContainer" containerID="2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.383522 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c\": container with ID starting with 2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c not found: ID does not exist" containerID="2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.383555 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c"} err="failed to get container status \"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c\": rpc error: code = NotFound desc = could not find container \"2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c\": container with ID starting with 2eea17853228eeb9d73886119958e0d43318edc482bebcfb91106b092ad7807c not found: ID does not exist" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.383593 4867 scope.go:117] "RemoveContainer" containerID="d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.384482 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6\": container with ID starting with d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6 not found: ID does not exist" containerID="d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.384512 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6"} err="failed to get container status \"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6\": rpc error: code = NotFound desc = could not find container \"d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6\": container with ID starting with d59275f19fd03f357a763888a40d179cd0f60bcb1c4f532b7b322fe13d3a8dd6 not found: ID does not exist" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.384532 4867 scope.go:117] "RemoveContainer" containerID="d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf" Dec 12 08:29:27 crc kubenswrapper[4867]: E1212 08:29:27.393458 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf\": container with ID starting with d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf not found: ID does not exist" containerID="d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.393503 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf"} err="failed to get container status \"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf\": rpc error: code = NotFound desc = could not find container \"d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf\": container with ID starting with d80a4823f1c9118187a671c309a97f0efb83941ebc2da8555a7b4db8c4e541bf not found: ID does not exist" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.394657 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "42806f5a-b1ec-47a6-b55e-494f7bfa8108" (UID: "42806f5a-b1ec-47a6-b55e-494f7bfa8108"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.397199 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9ff85fcb-63a4-4e7d-a22d-5ab4aaedf75b\") pod \"prometheus-metric-storage-0\" (UID: \"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e\") " pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.402305 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42806f5a-b1ec-47a6-b55e-494f7bfa8108" (UID: "42806f5a-b1ec-47a6-b55e-494f7bfa8108"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.447409 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "42806f5a-b1ec-47a6-b55e-494f7bfa8108" (UID: "42806f5a-b1ec-47a6-b55e-494f7bfa8108"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.467556 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd5fc\" (UniqueName: \"kubernetes.io/projected/42806f5a-b1ec-47a6-b55e-494f7bfa8108-kube-api-access-jd5fc\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.467596 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.467606 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.467618 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42806f5a-b1ec-47a6-b55e-494f7bfa8108-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.473923 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.956690 4867 scope.go:117] "RemoveContainer" containerID="ed1850cfd960e8db4ec02f08d50e9b23ce0f12d99a0428a1025b86a9a22517d8" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.956719 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.960926 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerStarted","Data":"38f991f5c6899d5b6dff1af1afdb519588775c6963cbd4a366c5c5457fa65c4b"} Dec 12 08:29:27 crc kubenswrapper[4867]: I1212 08:29:27.989867 4867 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" podUID="07f2128b-4e78-430a-9af7-45cf3e402573" Dec 12 08:29:28 crc kubenswrapper[4867]: I1212 08:29:28.036575 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 12 08:29:28 crc kubenswrapper[4867]: W1212 08:29:28.037315 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d045ee8_6266_4f6f_be38_0a8d39d1fb7e.slice/crio-93de1a377157ab37a9e4bf75460939e927222111098aaeccd627c17719caf4b7 WatchSource:0}: Error finding container 93de1a377157ab37a9e4bf75460939e927222111098aaeccd627c17719caf4b7: Status 404 returned error can't find the container with id 93de1a377157ab37a9e4bf75460939e927222111098aaeccd627c17719caf4b7 Dec 12 08:29:28 crc kubenswrapper[4867]: I1212 08:29:28.850583 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42806f5a-b1ec-47a6-b55e-494f7bfa8108" path="/var/lib/kubelet/pods/42806f5a-b1ec-47a6-b55e-494f7bfa8108/volumes" Dec 12 08:29:28 crc kubenswrapper[4867]: I1212 08:29:28.851744 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="815df45e-d1b3-4d94-9054-0ccc8f427af4" path="/var/lib/kubelet/pods/815df45e-d1b3-4d94-9054-0ccc8f427af4/volumes" Dec 12 08:29:28 crc kubenswrapper[4867]: I1212 08:29:28.971925 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerStarted","Data":"93de1a377157ab37a9e4bf75460939e927222111098aaeccd627c17719caf4b7"} Dec 12 08:29:29 crc kubenswrapper[4867]: I1212 08:29:29.986149 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerStarted","Data":"ed0bdfc1db1e7b7c6daeaccac97358b6a5206a67ce22d2c6069df86eb49fa52f"} Dec 12 08:29:29 crc kubenswrapper[4867]: I1212 08:29:29.986552 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 08:29:30 crc kubenswrapper[4867]: I1212 08:29:30.020249 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.648750614 podStartE2EDuration="13.020212501s" podCreationTimestamp="2025-12-12 08:29:17 +0000 UTC" firstStartedPulling="2025-12-12 08:29:18.680439279 +0000 UTC m=+6046.251820548" lastFinishedPulling="2025-12-12 08:29:29.051901146 +0000 UTC m=+6056.623282435" observedRunningTime="2025-12-12 08:29:30.007437024 +0000 UTC m=+6057.578818293" watchObservedRunningTime="2025-12-12 08:29:30.020212501 +0000 UTC m=+6057.591593770" Dec 12 08:29:32 crc kubenswrapper[4867]: I1212 08:29:32.024539 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerStarted","Data":"6050a5b290ebb3a994c477cedaa23a0af8782d36fd227b76d737c5b4b08a7c87"} Dec 12 08:29:33 crc kubenswrapper[4867]: I1212 08:29:33.840373 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:29:33 crc kubenswrapper[4867]: E1212 08:29:33.841034 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.352164 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-m6nfp"] Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.354498 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.389772 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-m6nfp"] Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.457665 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-5322-account-create-update-bql4k"] Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.459944 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.462746 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.468748 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5322-account-create-update-bql4k"] Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.493796 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmcx\" (UniqueName: \"kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.494011 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.595396 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwqmb\" (UniqueName: \"kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.595587 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.595622 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmcx\" (UniqueName: \"kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.595656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.596395 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.623845 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmcx\" (UniqueName: \"kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx\") pod \"aodh-db-create-m6nfp\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.689732 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.698798 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.698898 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwqmb\" (UniqueName: \"kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.699533 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.721003 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwqmb\" (UniqueName: \"kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb\") pod \"aodh-5322-account-create-update-bql4k\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:37 crc kubenswrapper[4867]: I1212 08:29:37.785011 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:38 crc kubenswrapper[4867]: I1212 08:29:38.087079 4867 generic.go:334] "Generic (PLEG): container finished" podID="6d045ee8-6266-4f6f-be38-0a8d39d1fb7e" containerID="6050a5b290ebb3a994c477cedaa23a0af8782d36fd227b76d737c5b4b08a7c87" exitCode=0 Dec 12 08:29:38 crc kubenswrapper[4867]: I1212 08:29:38.087124 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerDied","Data":"6050a5b290ebb3a994c477cedaa23a0af8782d36fd227b76d737c5b4b08a7c87"} Dec 12 08:29:38 crc kubenswrapper[4867]: I1212 08:29:38.196970 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-m6nfp"] Dec 12 08:29:38 crc kubenswrapper[4867]: I1212 08:29:38.344367 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-5322-account-create-update-bql4k"] Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.098965 4867 generic.go:334] "Generic (PLEG): container finished" podID="c258b557-6d4e-4ba6-b356-1a716ca6cd3b" containerID="46d38b6eb0e540ba5fbd73f8bf7407d6c2ddfdcee85ad1e0d27374e681efa3a5" exitCode=0 Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.099308 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5322-account-create-update-bql4k" event={"ID":"c258b557-6d4e-4ba6-b356-1a716ca6cd3b","Type":"ContainerDied","Data":"46d38b6eb0e540ba5fbd73f8bf7407d6c2ddfdcee85ad1e0d27374e681efa3a5"} Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.099433 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5322-account-create-update-bql4k" event={"ID":"c258b557-6d4e-4ba6-b356-1a716ca6cd3b","Type":"ContainerStarted","Data":"df05c950deeae09aefd707afeb07b3fb795e847d86ec2f2b60162f5469dc614f"} Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.101855 4867 generic.go:334] "Generic (PLEG): container finished" podID="0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" containerID="b4f4de8aa84fe727eade28ae8bfa7516b2a12a2038800ef6d5415ea12c79db54" exitCode=0 Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.101946 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m6nfp" event={"ID":"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93","Type":"ContainerDied","Data":"b4f4de8aa84fe727eade28ae8bfa7516b2a12a2038800ef6d5415ea12c79db54"} Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.101972 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m6nfp" event={"ID":"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93","Type":"ContainerStarted","Data":"fff90077e4c1a54222ab733119cdc2cf23d23950e0dde39644265036c7478da5"} Dec 12 08:29:39 crc kubenswrapper[4867]: I1212 08:29:39.103837 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerStarted","Data":"29a67d7b5ad12aaf55e9d126bfc7e901a3260d75881795d57a3a256aa45b7019"} Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.818000 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.824001 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.971063 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwqmb\" (UniqueName: \"kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb\") pod \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.971439 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vmcx\" (UniqueName: \"kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx\") pod \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.971518 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts\") pod \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\" (UID: \"c258b557-6d4e-4ba6-b356-1a716ca6cd3b\") " Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.971635 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts\") pod \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\" (UID: \"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93\") " Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.972171 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c258b557-6d4e-4ba6-b356-1a716ca6cd3b" (UID: "c258b557-6d4e-4ba6-b356-1a716ca6cd3b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.972267 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" (UID: "0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.974752 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.975265 4867 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.990896 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx" (OuterVolumeSpecName: "kube-api-access-4vmcx") pod "0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" (UID: "0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93"). InnerVolumeSpecName "kube-api-access-4vmcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:40 crc kubenswrapper[4867]: I1212 08:29:40.994868 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb" (OuterVolumeSpecName: "kube-api-access-rwqmb") pod "c258b557-6d4e-4ba6-b356-1a716ca6cd3b" (UID: "c258b557-6d4e-4ba6-b356-1a716ca6cd3b"). InnerVolumeSpecName "kube-api-access-rwqmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.077018 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwqmb\" (UniqueName: \"kubernetes.io/projected/c258b557-6d4e-4ba6-b356-1a716ca6cd3b-kube-api-access-rwqmb\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.077049 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vmcx\" (UniqueName: \"kubernetes.io/projected/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93-kube-api-access-4vmcx\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.120909 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-m6nfp" event={"ID":"0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93","Type":"ContainerDied","Data":"fff90077e4c1a54222ab733119cdc2cf23d23950e0dde39644265036c7478da5"} Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.120951 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fff90077e4c1a54222ab733119cdc2cf23d23950e0dde39644265036c7478da5" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.121006 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-m6nfp" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.122699 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-5322-account-create-update-bql4k" event={"ID":"c258b557-6d4e-4ba6-b356-1a716ca6cd3b","Type":"ContainerDied","Data":"df05c950deeae09aefd707afeb07b3fb795e847d86ec2f2b60162f5469dc614f"} Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.122725 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df05c950deeae09aefd707afeb07b3fb795e847d86ec2f2b60162f5469dc614f" Dec 12 08:29:41 crc kubenswrapper[4867]: I1212 08:29:41.122756 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-5322-account-create-update-bql4k" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.136483 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerStarted","Data":"ec0dd7851ccf14f159a2721cbf6ed69d36347be50fc53a1b10898c9c40f229ff"} Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.136812 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"6d045ee8-6266-4f6f-be38-0a8d39d1fb7e","Type":"ContainerStarted","Data":"d59a66f25d1b0a2f90bc393163176499edb33f4dbd08ae526d4c0be24edb27af"} Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.166084 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.166058585 podStartE2EDuration="16.166058585s" podCreationTimestamp="2025-12-12 08:29:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:29:42.157745829 +0000 UTC m=+6069.729127098" watchObservedRunningTime="2025-12-12 08:29:42.166058585 +0000 UTC m=+6069.737439844" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.474171 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.474332 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.479400 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.782240 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-8bq7b"] Dec 12 08:29:42 crc kubenswrapper[4867]: E1212 08:29:42.782693 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" containerName="mariadb-database-create" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.782715 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" containerName="mariadb-database-create" Dec 12 08:29:42 crc kubenswrapper[4867]: E1212 08:29:42.782748 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c258b557-6d4e-4ba6-b356-1a716ca6cd3b" containerName="mariadb-account-create-update" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.782756 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c258b557-6d4e-4ba6-b356-1a716ca6cd3b" containerName="mariadb-account-create-update" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.782946 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" containerName="mariadb-database-create" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.782967 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c258b557-6d4e-4ba6-b356-1a716ca6cd3b" containerName="mariadb-account-create-update" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.783637 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.785739 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.786781 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-fd7dv" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.786974 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.787159 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.798998 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-8bq7b"] Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.913411 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.913823 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.914118 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:42 crc kubenswrapper[4867]: I1212 08:29:42.914174 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gftd\" (UniqueName: \"kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.016028 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.016126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.017296 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.017353 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gftd\" (UniqueName: \"kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.021965 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.022140 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.023201 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.038071 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gftd\" (UniqueName: \"kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd\") pod \"aodh-db-sync-8bq7b\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.104510 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.150394 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 12 08:29:43 crc kubenswrapper[4867]: W1212 08:29:43.633303 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39a39479_d1b6_45a2_a480_c15c61be7627.slice/crio-a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0 WatchSource:0}: Error finding container a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0: Status 404 returned error can't find the container with id a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0 Dec 12 08:29:43 crc kubenswrapper[4867]: I1212 08:29:43.633975 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-8bq7b"] Dec 12 08:29:44 crc kubenswrapper[4867]: I1212 08:29:44.160241 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8bq7b" event={"ID":"39a39479-d1b6-45a2-a480-c15c61be7627","Type":"ContainerStarted","Data":"a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0"} Dec 12 08:29:47 crc kubenswrapper[4867]: I1212 08:29:47.840712 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:29:47 crc kubenswrapper[4867]: E1212 08:29:47.841466 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:29:48 crc kubenswrapper[4867]: I1212 08:29:48.042774 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 12 08:29:49 crc kubenswrapper[4867]: I1212 08:29:49.219277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8bq7b" event={"ID":"39a39479-d1b6-45a2-a480-c15c61be7627","Type":"ContainerStarted","Data":"e952512f990fb58145146f8d3812dc427af9c69fe0b57c73ca6e72026b02d963"} Dec 12 08:29:49 crc kubenswrapper[4867]: I1212 08:29:49.237679 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-8bq7b" podStartSLOduration=2.199511635 podStartE2EDuration="7.237652333s" podCreationTimestamp="2025-12-12 08:29:42 +0000 UTC" firstStartedPulling="2025-12-12 08:29:43.635573209 +0000 UTC m=+6071.206954468" lastFinishedPulling="2025-12-12 08:29:48.673713897 +0000 UTC m=+6076.245095166" observedRunningTime="2025-12-12 08:29:49.23509174 +0000 UTC m=+6076.806472999" watchObservedRunningTime="2025-12-12 08:29:49.237652333 +0000 UTC m=+6076.809033602" Dec 12 08:29:51 crc kubenswrapper[4867]: I1212 08:29:51.780007 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:51 crc kubenswrapper[4867]: I1212 08:29:51.780562 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" containerName="kube-state-metrics" containerID="cri-o://dcf41a6869ed2e886079b80edbb9d4a3f839e14309b6092fa8811d6f95c5e8c0" gracePeriod=30 Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.261064 4867 generic.go:334] "Generic (PLEG): container finished" podID="39a39479-d1b6-45a2-a480-c15c61be7627" containerID="e952512f990fb58145146f8d3812dc427af9c69fe0b57c73ca6e72026b02d963" exitCode=0 Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.261119 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8bq7b" event={"ID":"39a39479-d1b6-45a2-a480-c15c61be7627","Type":"ContainerDied","Data":"e952512f990fb58145146f8d3812dc427af9c69fe0b57c73ca6e72026b02d963"} Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.265692 4867 generic.go:334] "Generic (PLEG): container finished" podID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" containerID="dcf41a6869ed2e886079b80edbb9d4a3f839e14309b6092fa8811d6f95c5e8c0" exitCode=2 Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.265733 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d","Type":"ContainerDied","Data":"dcf41a6869ed2e886079b80edbb9d4a3f839e14309b6092fa8811d6f95c5e8c0"} Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.265759 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d","Type":"ContainerDied","Data":"81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb"} Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.265770 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81987ae1c951c186549d6ac67702252ae07dd2df56741303c78218a3879435fb" Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.329614 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.424253 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkssn\" (UniqueName: \"kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn\") pod \"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d\" (UID: \"4d97ed7f-b8eb-4a0e-983d-71d9aa20648d\") " Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.433400 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn" (OuterVolumeSpecName: "kube-api-access-vkssn") pod "4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" (UID: "4d97ed7f-b8eb-4a0e-983d-71d9aa20648d"). InnerVolumeSpecName "kube-api-access-vkssn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:52 crc kubenswrapper[4867]: I1212 08:29:52.527162 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkssn\" (UniqueName: \"kubernetes.io/projected/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d-kube-api-access-vkssn\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.275583 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.317146 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.335699 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.347157 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:53 crc kubenswrapper[4867]: E1212 08:29:53.350624 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" containerName="kube-state-metrics" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.350668 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" containerName="kube-state-metrics" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.351078 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" containerName="kube-state-metrics" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.352021 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.355049 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.355266 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.371274 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.445849 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.446044 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.446107 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.446151 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx44c\" (UniqueName: \"kubernetes.io/projected/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-api-access-zx44c\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.548675 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.548756 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.548830 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx44c\" (UniqueName: \"kubernetes.io/projected/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-api-access-zx44c\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.548911 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.554484 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.555562 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.555912 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a7fe15-90d4-483c-b893-5c69e8a969da-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.565162 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx44c\" (UniqueName: \"kubernetes.io/projected/94a7fe15-90d4-483c-b893-5c69e8a969da-kube-api-access-zx44c\") pod \"kube-state-metrics-0\" (UID: \"94a7fe15-90d4-483c-b893-5c69e8a969da\") " pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.660578 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.675496 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.752984 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gftd\" (UniqueName: \"kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd\") pod \"39a39479-d1b6-45a2-a480-c15c61be7627\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.753040 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data\") pod \"39a39479-d1b6-45a2-a480-c15c61be7627\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.753056 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts\") pod \"39a39479-d1b6-45a2-a480-c15c61be7627\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.753086 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle\") pod \"39a39479-d1b6-45a2-a480-c15c61be7627\" (UID: \"39a39479-d1b6-45a2-a480-c15c61be7627\") " Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.757994 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd" (OuterVolumeSpecName: "kube-api-access-7gftd") pod "39a39479-d1b6-45a2-a480-c15c61be7627" (UID: "39a39479-d1b6-45a2-a480-c15c61be7627"). InnerVolumeSpecName "kube-api-access-7gftd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.758102 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts" (OuterVolumeSpecName: "scripts") pod "39a39479-d1b6-45a2-a480-c15c61be7627" (UID: "39a39479-d1b6-45a2-a480-c15c61be7627"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.794664 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.794966 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-central-agent" containerID="cri-o://303fec35adcf223befd5fb8ce60bcf535b8bda87ed35edd1d2da82fc96a72db6" gracePeriod=30 Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.795099 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="proxy-httpd" containerID="cri-o://ed0bdfc1db1e7b7c6daeaccac97358b6a5206a67ce22d2c6069df86eb49fa52f" gracePeriod=30 Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.795157 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="sg-core" containerID="cri-o://38f991f5c6899d5b6dff1af1afdb519588775c6963cbd4a366c5c5457fa65c4b" gracePeriod=30 Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.795200 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-notification-agent" containerID="cri-o://c1d30a17c8ae74d8ff1b6da24997ec31f28503af3d802503cd87a92421f710eb" gracePeriod=30 Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.801499 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39a39479-d1b6-45a2-a480-c15c61be7627" (UID: "39a39479-d1b6-45a2-a480-c15c61be7627"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.813580 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data" (OuterVolumeSpecName: "config-data") pod "39a39479-d1b6-45a2-a480-c15c61be7627" (UID: "39a39479-d1b6-45a2-a480-c15c61be7627"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.857927 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gftd\" (UniqueName: \"kubernetes.io/projected/39a39479-d1b6-45a2-a480-c15c61be7627-kube-api-access-7gftd\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.857984 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.857994 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:53 crc kubenswrapper[4867]: I1212 08:29:53.858002 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a39479-d1b6-45a2-a480-c15c61be7627-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.176381 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 12 08:29:54 crc kubenswrapper[4867]: W1212 08:29:54.179067 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a7fe15_90d4_483c_b893_5c69e8a969da.slice/crio-11a9f3a8ffe53896ab4f81c80b057733bc66d5c4ab9566a22a7772f68bdb5633 WatchSource:0}: Error finding container 11a9f3a8ffe53896ab4f81c80b057733bc66d5c4ab9566a22a7772f68bdb5633: Status 404 returned error can't find the container with id 11a9f3a8ffe53896ab4f81c80b057733bc66d5c4ab9566a22a7772f68bdb5633 Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.288183 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8bq7b" event={"ID":"39a39479-d1b6-45a2-a480-c15c61be7627","Type":"ContainerDied","Data":"a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0"} Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.288242 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b45fc045bd6a1fe72db4e97568e99f37802552607b32f40d68f69b245944e0" Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.288314 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8bq7b" Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.294669 4867 generic.go:334] "Generic (PLEG): container finished" podID="134c1312-8535-4138-9c01-d245f06010a4" containerID="ed0bdfc1db1e7b7c6daeaccac97358b6a5206a67ce22d2c6069df86eb49fa52f" exitCode=0 Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.294697 4867 generic.go:334] "Generic (PLEG): container finished" podID="134c1312-8535-4138-9c01-d245f06010a4" containerID="38f991f5c6899d5b6dff1af1afdb519588775c6963cbd4a366c5c5457fa65c4b" exitCode=2 Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.294769 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerDied","Data":"ed0bdfc1db1e7b7c6daeaccac97358b6a5206a67ce22d2c6069df86eb49fa52f"} Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.294852 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerDied","Data":"38f991f5c6899d5b6dff1af1afdb519588775c6963cbd4a366c5c5457fa65c4b"} Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.296361 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94a7fe15-90d4-483c-b893-5c69e8a969da","Type":"ContainerStarted","Data":"11a9f3a8ffe53896ab4f81c80b057733bc66d5c4ab9566a22a7772f68bdb5633"} Dec 12 08:29:54 crc kubenswrapper[4867]: I1212 08:29:54.851756 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d97ed7f-b8eb-4a0e-983d-71d9aa20648d" path="/var/lib/kubelet/pods/4d97ed7f-b8eb-4a0e-983d-71d9aa20648d/volumes" Dec 12 08:29:55 crc kubenswrapper[4867]: I1212 08:29:55.311190 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"94a7fe15-90d4-483c-b893-5c69e8a969da","Type":"ContainerStarted","Data":"9a2948b27a088e8da95b777e55271d3e499dc0d3c853c53666cd191a56bc79f8"} Dec 12 08:29:55 crc kubenswrapper[4867]: I1212 08:29:55.311711 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 12 08:29:55 crc kubenswrapper[4867]: I1212 08:29:55.318091 4867 generic.go:334] "Generic (PLEG): container finished" podID="134c1312-8535-4138-9c01-d245f06010a4" containerID="303fec35adcf223befd5fb8ce60bcf535b8bda87ed35edd1d2da82fc96a72db6" exitCode=0 Dec 12 08:29:55 crc kubenswrapper[4867]: I1212 08:29:55.318139 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerDied","Data":"303fec35adcf223befd5fb8ce60bcf535b8bda87ed35edd1d2da82fc96a72db6"} Dec 12 08:29:55 crc kubenswrapper[4867]: I1212 08:29:55.335886 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.97769265 podStartE2EDuration="2.335866333s" podCreationTimestamp="2025-12-12 08:29:53 +0000 UTC" firstStartedPulling="2025-12-12 08:29:54.181301869 +0000 UTC m=+6081.752683138" lastFinishedPulling="2025-12-12 08:29:54.539475552 +0000 UTC m=+6082.110856821" observedRunningTime="2025-12-12 08:29:55.329590417 +0000 UTC m=+6082.900971686" watchObservedRunningTime="2025-12-12 08:29:55.335866333 +0000 UTC m=+6082.907247602" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.325374 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 12 08:29:57 crc kubenswrapper[4867]: E1212 08:29:57.326958 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a39479-d1b6-45a2-a480-c15c61be7627" containerName="aodh-db-sync" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.327019 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a39479-d1b6-45a2-a480-c15c61be7627" containerName="aodh-db-sync" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.327296 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="39a39479-d1b6-45a2-a480-c15c61be7627" containerName="aodh-db-sync" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.330380 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.334664 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.334875 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-fd7dv" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.335080 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.387672 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.436931 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67k49\" (UniqueName: \"kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.436993 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.437029 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.437046 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.538947 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67k49\" (UniqueName: \"kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.539002 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.539036 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.539052 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.548756 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.554107 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.554493 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.561703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67k49\" (UniqueName: \"kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49\") pod \"aodh-0\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " pod="openstack/aodh-0" Dec 12 08:29:57 crc kubenswrapper[4867]: I1212 08:29:57.672366 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:29:58 crc kubenswrapper[4867]: W1212 08:29:58.213691 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c WatchSource:0}: Error finding container f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c: Status 404 returned error can't find the container with id f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.214358 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.366835 4867 generic.go:334] "Generic (PLEG): container finished" podID="134c1312-8535-4138-9c01-d245f06010a4" containerID="c1d30a17c8ae74d8ff1b6da24997ec31f28503af3d802503cd87a92421f710eb" exitCode=0 Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.366925 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerDied","Data":"c1d30a17c8ae74d8ff1b6da24997ec31f28503af3d802503cd87a92421f710eb"} Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.368630 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerStarted","Data":"f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c"} Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.690909 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.866733 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.866856 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.866947 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hbb4\" (UniqueName: \"kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.866966 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.866980 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.867071 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.867095 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts\") pod \"134c1312-8535-4138-9c01-d245f06010a4\" (UID: \"134c1312-8535-4138-9c01-d245f06010a4\") " Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.869926 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.870017 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.879744 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4" (OuterVolumeSpecName: "kube-api-access-6hbb4") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "kube-api-access-6hbb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.900641 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts" (OuterVolumeSpecName: "scripts") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.912209 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.977864 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.978277 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hbb4\" (UniqueName: \"kubernetes.io/projected/134c1312-8535-4138-9c01-d245f06010a4-kube-api-access-6hbb4\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.978291 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/134c1312-8535-4138-9c01-d245f06010a4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.978299 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:58 crc kubenswrapper[4867]: I1212 08:29:58.978308 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:58.996961 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.039753 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data" (OuterVolumeSpecName: "config-data") pod "134c1312-8535-4138-9c01-d245f06010a4" (UID: "134c1312-8535-4138-9c01-d245f06010a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.080510 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.080545 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/134c1312-8535-4138-9c01-d245f06010a4-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.282056 4867 scope.go:117] "RemoveContainer" containerID="527e851f4b1a85e49e4d67716cc580ab031c8f6e47294d3515ba32dccc5f94a8" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.312772 4867 scope.go:117] "RemoveContainer" containerID="18fc7f62366a413ad09766fb175f37064fcd6f4998d7d3dc26af29e6825a741d" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.398981 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.398997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"134c1312-8535-4138-9c01-d245f06010a4","Type":"ContainerDied","Data":"19ed42d4d724614fe348df2b0c45dbd9a9b66414095b8903b923f94876560f24"} Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.399064 4867 scope.go:117] "RemoveContainer" containerID="ed0bdfc1db1e7b7c6daeaccac97358b6a5206a67ce22d2c6069df86eb49fa52f" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.403945 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerStarted","Data":"da1f4fdebef1710c708a97ac47fbc61bf235631de9f2574c29010358b9591a26"} Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.466387 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.483774 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.515294 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:59 crc kubenswrapper[4867]: E1212 08:29:59.515714 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="sg-core" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.515733 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="sg-core" Dec 12 08:29:59 crc kubenswrapper[4867]: E1212 08:29:59.515761 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="proxy-httpd" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.515768 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="proxy-httpd" Dec 12 08:29:59 crc kubenswrapper[4867]: E1212 08:29:59.515783 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-notification-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.515789 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-notification-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: E1212 08:29:59.515799 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-central-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.515805 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-central-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.516004 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="sg-core" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.516024 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-notification-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.516038 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="proxy-httpd" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.516051 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="134c1312-8535-4138-9c01-d245f06010a4" containerName="ceilometer-central-agent" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.518260 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.524350 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.524886 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.525184 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.525379 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.690878 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.690934 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.690965 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.691018 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.691049 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.691081 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hjf2\" (UniqueName: \"kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.691103 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.691160 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.782352 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:29:59 crc kubenswrapper[4867]: E1212 08:29:59.783434 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-9hjf2 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="4caef63d-3dbd-4a34-baa1-dc9e946babb3" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792587 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792647 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792680 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792700 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792731 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792760 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hjf2\" (UniqueName: \"kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792781 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.792829 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.793454 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.793532 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.798382 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.799191 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.799465 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.808340 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.810280 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.823769 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hjf2\" (UniqueName: \"kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2\") pod \"ceilometer-0\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " pod="openstack/ceilometer-0" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.908436 4867 scope.go:117] "RemoveContainer" containerID="38f991f5c6899d5b6dff1af1afdb519588775c6963cbd4a366c5c5457fa65c4b" Dec 12 08:29:59 crc kubenswrapper[4867]: I1212 08:29:59.946213 4867 scope.go:117] "RemoveContainer" containerID="0af4f965ee49f7ce8411b711a483551419065977c7b1ab1fa73d8f315ebe78ba" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.016356 4867 scope.go:117] "RemoveContainer" containerID="c1d30a17c8ae74d8ff1b6da24997ec31f28503af3d802503cd87a92421f710eb" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.074649 4867 scope.go:117] "RemoveContainer" containerID="303fec35adcf223befd5fb8ce60bcf535b8bda87ed35edd1d2da82fc96a72db6" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.149538 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff"] Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.151384 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.156643 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.156700 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.167169 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff"] Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.302737 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99nc7\" (UniqueName: \"kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.302836 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.303044 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.405668 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.406055 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99nc7\" (UniqueName: \"kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.406109 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.406707 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.411352 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.424913 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerStarted","Data":"cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f"} Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.424931 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.434075 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99nc7\" (UniqueName: \"kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7\") pod \"collect-profiles-29425470-dzdff\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.436536 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.472852 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609715 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609753 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609828 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609862 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609886 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hjf2\" (UniqueName: \"kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.609907 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.610152 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.610156 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.610186 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd\") pod \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\" (UID: \"4caef63d-3dbd-4a34-baa1-dc9e946babb3\") " Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.610622 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.610986 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.615469 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2" (OuterVolumeSpecName: "kube-api-access-9hjf2") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "kube-api-access-9hjf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.617976 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.618077 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts" (OuterVolumeSpecName: "scripts") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.618211 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data" (OuterVolumeSpecName: "config-data") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.618353 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.620643 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4caef63d-3dbd-4a34-baa1-dc9e946babb3" (UID: "4caef63d-3dbd-4a34-baa1-dc9e946babb3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714074 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714461 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4caef63d-3dbd-4a34-baa1-dc9e946babb3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714472 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714480 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714491 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714503 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hjf2\" (UniqueName: \"kubernetes.io/projected/4caef63d-3dbd-4a34-baa1-dc9e946babb3-kube-api-access-9hjf2\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.714512 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4caef63d-3dbd-4a34-baa1-dc9e946babb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.840849 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:30:00 crc kubenswrapper[4867]: E1212 08:30:00.841219 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:30:00 crc kubenswrapper[4867]: I1212 08:30:00.855512 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="134c1312-8535-4138-9c01-d245f06010a4" path="/var/lib/kubelet/pods/134c1312-8535-4138-9c01-d245f06010a4/volumes" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.023110 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.236728 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.440271 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.441507 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" event={"ID":"83f1ab72-a940-4d62-a67c-367662591295","Type":"ContainerStarted","Data":"99b5596959fc5842036d24edfe78ffef6eed4e790cead83e4dba8aa177600f84"} Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.441540 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" event={"ID":"83f1ab72-a940-4d62-a67c-367662591295","Type":"ContainerStarted","Data":"a02079fee64e2f13e7238ce63a7f5e825b29338075bebab2668557656f10d04a"} Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.469127 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" podStartSLOduration=1.4691039799999999 podStartE2EDuration="1.46910398s" podCreationTimestamp="2025-12-12 08:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:30:01.463209364 +0000 UTC m=+6089.034590653" watchObservedRunningTime="2025-12-12 08:30:01.46910398 +0000 UTC m=+6089.040485249" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.516924 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.529320 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.548048 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.554625 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.560183 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.564637 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.564848 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.603713 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737037 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737112 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737145 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737441 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737578 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.737803 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.738015 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.738862 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gnx8\" (UniqueName: \"kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.840484 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.840875 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gnx8\" (UniqueName: \"kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.840919 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.840978 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.841023 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.841056 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.841097 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.841170 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.841664 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.842415 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.844530 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.845161 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.845849 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.845890 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.848652 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.859836 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gnx8\" (UniqueName: \"kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8\") pod \"ceilometer-0\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " pod="openstack/ceilometer-0" Dec 12 08:30:01 crc kubenswrapper[4867]: I1212 08:30:01.879333 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.372272 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:02 crc kubenswrapper[4867]: W1212 08:30:02.383748 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a0693b8_b55e_42a7_bfdb_8d3778b89f7c.slice/crio-b981cee93f8d7a795bb68277d9dc83261009b90682783a1cef56bc3a0a731542 WatchSource:0}: Error finding container b981cee93f8d7a795bb68277d9dc83261009b90682783a1cef56bc3a0a731542: Status 404 returned error can't find the container with id b981cee93f8d7a795bb68277d9dc83261009b90682783a1cef56bc3a0a731542 Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.471255 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerStarted","Data":"b981cee93f8d7a795bb68277d9dc83261009b90682783a1cef56bc3a0a731542"} Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.482916 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerStarted","Data":"3dd92f81cf3cb450c7ef7631a2bcbd0f7a688dab5775df28bdaed7788b385d33"} Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.493095 4867 generic.go:334] "Generic (PLEG): container finished" podID="83f1ab72-a940-4d62-a67c-367662591295" containerID="99b5596959fc5842036d24edfe78ffef6eed4e790cead83e4dba8aa177600f84" exitCode=0 Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.493157 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" event={"ID":"83f1ab72-a940-4d62-a67c-367662591295","Type":"ContainerDied","Data":"99b5596959fc5842036d24edfe78ffef6eed4e790cead83e4dba8aa177600f84"} Dec 12 08:30:02 crc kubenswrapper[4867]: I1212 08:30:02.855718 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4caef63d-3dbd-4a34-baa1-dc9e946babb3" path="/var/lib/kubelet/pods/4caef63d-3dbd-4a34-baa1-dc9e946babb3/volumes" Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.504104 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerStarted","Data":"3a1c3b79dd633c2960ba54a2dd6b22e05d0640a8c30ff9353c5b108cb40e936a"} Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.702948 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.821552 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.994376 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume\") pod \"83f1ab72-a940-4d62-a67c-367662591295\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.994776 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99nc7\" (UniqueName: \"kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7\") pod \"83f1ab72-a940-4d62-a67c-367662591295\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.994810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume\") pod \"83f1ab72-a940-4d62-a67c-367662591295\" (UID: \"83f1ab72-a940-4d62-a67c-367662591295\") " Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.995811 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume" (OuterVolumeSpecName: "config-volume") pod "83f1ab72-a940-4d62-a67c-367662591295" (UID: "83f1ab72-a940-4d62-a67c-367662591295"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:30:03 crc kubenswrapper[4867]: I1212 08:30:03.999785 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83f1ab72-a940-4d62-a67c-367662591295" (UID: "83f1ab72-a940-4d62-a67c-367662591295"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.002546 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7" (OuterVolumeSpecName: "kube-api-access-99nc7") pod "83f1ab72-a940-4d62-a67c-367662591295" (UID: "83f1ab72-a940-4d62-a67c-367662591295"). InnerVolumeSpecName "kube-api-access-99nc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.097296 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f1ab72-a940-4d62-a67c-367662591295-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.097339 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f1ab72-a940-4d62-a67c-367662591295-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.097349 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99nc7\" (UniqueName: \"kubernetes.io/projected/83f1ab72-a940-4d62-a67c-367662591295-kube-api-access-99nc7\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.491590 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.515486 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerStarted","Data":"2521d068d6818a80a12fd4e6471a955013d8fc1c55c236c60dedb1b10cd46224"} Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.515532 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerStarted","Data":"46d0fc9d73df7022831d3764ec51c4b2cc6546b4613cb379f789bc779c559340"} Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.518043 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerStarted","Data":"bb4c283c6bd370853f5115b370aca1fb9435eae29ee8b3767ee4abb434ad342d"} Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.518203 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-listener" containerID="cri-o://bb4c283c6bd370853f5115b370aca1fb9435eae29ee8b3767ee4abb434ad342d" gracePeriod=30 Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.518345 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-notifier" containerID="cri-o://3dd92f81cf3cb450c7ef7631a2bcbd0f7a688dab5775df28bdaed7788b385d33" gracePeriod=30 Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.518382 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-evaluator" containerID="cri-o://cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f" gracePeriod=30 Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.518148 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-api" containerID="cri-o://da1f4fdebef1710c708a97ac47fbc61bf235631de9f2574c29010358b9591a26" gracePeriod=30 Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.520399 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" event={"ID":"83f1ab72-a940-4d62-a67c-367662591295","Type":"ContainerDied","Data":"a02079fee64e2f13e7238ce63a7f5e825b29338075bebab2668557656f10d04a"} Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.520429 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a02079fee64e2f13e7238ce63a7f5e825b29338075bebab2668557656f10d04a" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.520485 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.556467 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg"] Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.572163 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425425-rpdcg"] Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.575487 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.590175393 podStartE2EDuration="7.57546579s" podCreationTimestamp="2025-12-12 08:29:57 +0000 UTC" firstStartedPulling="2025-12-12 08:29:58.217707874 +0000 UTC m=+6085.789089133" lastFinishedPulling="2025-12-12 08:30:03.202998261 +0000 UTC m=+6090.774379530" observedRunningTime="2025-12-12 08:30:04.555198287 +0000 UTC m=+6092.126579556" watchObservedRunningTime="2025-12-12 08:30:04.57546579 +0000 UTC m=+6092.146847049" Dec 12 08:30:04 crc kubenswrapper[4867]: I1212 08:30:04.892177 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1526b2-f850-4458-b1fa-b6f094eb2c26" path="/var/lib/kubelet/pods/8d1526b2-f850-4458-b1fa-b6f094eb2c26/volumes" Dec 12 08:30:05 crc kubenswrapper[4867]: E1212 08:30:05.105434 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-conmon-cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533255 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerID="3dd92f81cf3cb450c7ef7631a2bcbd0f7a688dab5775df28bdaed7788b385d33" exitCode=0 Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533595 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerID="cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f" exitCode=0 Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533609 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerID="da1f4fdebef1710c708a97ac47fbc61bf235631de9f2574c29010358b9591a26" exitCode=0 Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533320 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerDied","Data":"3dd92f81cf3cb450c7ef7631a2bcbd0f7a688dab5775df28bdaed7788b385d33"} Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533646 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerDied","Data":"cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f"} Dec 12 08:30:05 crc kubenswrapper[4867]: I1212 08:30:05.533659 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerDied","Data":"da1f4fdebef1710c708a97ac47fbc61bf235631de9f2574c29010358b9591a26"} Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544250 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerStarted","Data":"d2f224e36e1e7f2a13438e91f11e75f05431e99c99e016f55fb3444f6fe12eb8"} Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544579 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544388 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-central-agent" containerID="cri-o://3a1c3b79dd633c2960ba54a2dd6b22e05d0640a8c30ff9353c5b108cb40e936a" gracePeriod=30 Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544616 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="proxy-httpd" containerID="cri-o://d2f224e36e1e7f2a13438e91f11e75f05431e99c99e016f55fb3444f6fe12eb8" gracePeriod=30 Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544647 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-notification-agent" containerID="cri-o://46d0fc9d73df7022831d3764ec51c4b2cc6546b4613cb379f789bc779c559340" gracePeriod=30 Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.544672 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="sg-core" containerID="cri-o://2521d068d6818a80a12fd4e6471a955013d8fc1c55c236c60dedb1b10cd46224" gracePeriod=30 Dec 12 08:30:06 crc kubenswrapper[4867]: I1212 08:30:06.580939 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.736877722 podStartE2EDuration="5.580918846s" podCreationTimestamp="2025-12-12 08:30:01 +0000 UTC" firstStartedPulling="2025-12-12 08:30:02.38681168 +0000 UTC m=+6089.958192939" lastFinishedPulling="2025-12-12 08:30:06.230852794 +0000 UTC m=+6093.802234063" observedRunningTime="2025-12-12 08:30:06.567604066 +0000 UTC m=+6094.138985345" watchObservedRunningTime="2025-12-12 08:30:06.580918846 +0000 UTC m=+6094.152300115" Dec 12 08:30:07 crc kubenswrapper[4867]: I1212 08:30:07.555744 4867 generic.go:334] "Generic (PLEG): container finished" podID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerID="2521d068d6818a80a12fd4e6471a955013d8fc1c55c236c60dedb1b10cd46224" exitCode=2 Dec 12 08:30:07 crc kubenswrapper[4867]: I1212 08:30:07.556072 4867 generic.go:334] "Generic (PLEG): container finished" podID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerID="46d0fc9d73df7022831d3764ec51c4b2cc6546b4613cb379f789bc779c559340" exitCode=0 Dec 12 08:30:07 crc kubenswrapper[4867]: I1212 08:30:07.555821 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerDied","Data":"2521d068d6818a80a12fd4e6471a955013d8fc1c55c236c60dedb1b10cd46224"} Dec 12 08:30:07 crc kubenswrapper[4867]: I1212 08:30:07.556110 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerDied","Data":"46d0fc9d73df7022831d3764ec51c4b2cc6546b4613cb379f789bc779c559340"} Dec 12 08:30:11 crc kubenswrapper[4867]: I1212 08:30:11.596992 4867 generic.go:334] "Generic (PLEG): container finished" podID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerID="3a1c3b79dd633c2960ba54a2dd6b22e05d0640a8c30ff9353c5b108cb40e936a" exitCode=0 Dec 12 08:30:11 crc kubenswrapper[4867]: I1212 08:30:11.597105 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerDied","Data":"3a1c3b79dd633c2960ba54a2dd6b22e05d0640a8c30ff9353c5b108cb40e936a"} Dec 12 08:30:14 crc kubenswrapper[4867]: I1212 08:30:14.837793 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:30:14 crc kubenswrapper[4867]: E1212 08:30:14.838587 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:30:27 crc kubenswrapper[4867]: I1212 08:30:27.839169 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:30:27 crc kubenswrapper[4867]: E1212 08:30:27.840533 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:30:31 crc kubenswrapper[4867]: I1212 08:30:31.888764 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 12 08:30:34 crc kubenswrapper[4867]: I1212 08:30:34.816575 4867 generic.go:334] "Generic (PLEG): container finished" podID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerID="bb4c283c6bd370853f5115b370aca1fb9435eae29ee8b3767ee4abb434ad342d" exitCode=137 Dec 12 08:30:34 crc kubenswrapper[4867]: I1212 08:30:34.816634 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerDied","Data":"bb4c283c6bd370853f5115b370aca1fb9435eae29ee8b3767ee4abb434ad342d"} Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.415109 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.613965 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts\") pod \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.614184 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data\") pod \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.614294 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle\") pod \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.614352 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67k49\" (UniqueName: \"kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49\") pod \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\" (UID: \"5f762879-f3a1-47e2-a6fc-a60dc14701e4\") " Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.621649 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts" (OuterVolumeSpecName: "scripts") pod "5f762879-f3a1-47e2-a6fc-a60dc14701e4" (UID: "5f762879-f3a1-47e2-a6fc-a60dc14701e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.640451 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49" (OuterVolumeSpecName: "kube-api-access-67k49") pod "5f762879-f3a1-47e2-a6fc-a60dc14701e4" (UID: "5f762879-f3a1-47e2-a6fc-a60dc14701e4"). InnerVolumeSpecName "kube-api-access-67k49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.716154 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67k49\" (UniqueName: \"kubernetes.io/projected/5f762879-f3a1-47e2-a6fc-a60dc14701e4-kube-api-access-67k49\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.716193 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.740120 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data" (OuterVolumeSpecName: "config-data") pod "5f762879-f3a1-47e2-a6fc-a60dc14701e4" (UID: "5f762879-f3a1-47e2-a6fc-a60dc14701e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.744668 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f762879-f3a1-47e2-a6fc-a60dc14701e4" (UID: "5f762879-f3a1-47e2-a6fc-a60dc14701e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.818176 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.818207 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f762879-f3a1-47e2-a6fc-a60dc14701e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.829609 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"5f762879-f3a1-47e2-a6fc-a60dc14701e4","Type":"ContainerDied","Data":"f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c"} Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.829673 4867 scope.go:117] "RemoveContainer" containerID="bb4c283c6bd370853f5115b370aca1fb9435eae29ee8b3767ee4abb434ad342d" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.829682 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.857059 4867 scope.go:117] "RemoveContainer" containerID="3dd92f81cf3cb450c7ef7631a2bcbd0f7a688dab5775df28bdaed7788b385d33" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.877135 4867 scope.go:117] "RemoveContainer" containerID="cc45c0160debad83e05dade15961eb61be85ddefd66bb391d2474a778dbf042f" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.902009 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.922182 4867 scope.go:117] "RemoveContainer" containerID="da1f4fdebef1710c708a97ac47fbc61bf235631de9f2574c29010358b9591a26" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.922379 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.934038 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:35 crc kubenswrapper[4867]: E1212 08:30:35.934612 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-notifier" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.934684 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-notifier" Dec 12 08:30:35 crc kubenswrapper[4867]: E1212 08:30:35.934742 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-listener" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.934831 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-listener" Dec 12 08:30:35 crc kubenswrapper[4867]: E1212 08:30:35.934881 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-evaluator" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.934937 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-evaluator" Dec 12 08:30:35 crc kubenswrapper[4867]: E1212 08:30:35.935001 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-api" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935053 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-api" Dec 12 08:30:35 crc kubenswrapper[4867]: E1212 08:30:35.935123 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f1ab72-a940-4d62-a67c-367662591295" containerName="collect-profiles" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935188 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f1ab72-a940-4d62-a67c-367662591295" containerName="collect-profiles" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935493 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-api" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935605 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-evaluator" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935683 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-listener" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935764 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f1ab72-a940-4d62-a67c-367662591295" containerName="collect-profiles" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.935838 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" containerName="aodh-notifier" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.937944 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.941056 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.941364 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-fd7dv" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.941399 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.941510 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.942426 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 12 08:30:35 crc kubenswrapper[4867]: I1212 08:30:35.946304 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123105 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-internal-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123173 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123206 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-config-data\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123488 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-scripts\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123649 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqpc6\" (UniqueName: \"kubernetes.io/projected/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-kube-api-access-sqpc6\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.123732 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-public-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.225797 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-internal-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.225851 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.225874 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-config-data\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.225941 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-scripts\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.225974 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqpc6\" (UniqueName: \"kubernetes.io/projected/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-kube-api-access-sqpc6\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.226021 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-public-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.229440 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-scripts\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.232024 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-config-data\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.238759 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-internal-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.241194 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.245957 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqpc6\" (UniqueName: \"kubernetes.io/projected/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-kube-api-access-sqpc6\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.245991 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0b99c63-9b5a-4ffe-92f4-5fef96b3df72-public-tls-certs\") pod \"aodh-0\" (UID: \"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72\") " pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.255757 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.739978 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 12 08:30:36 crc kubenswrapper[4867]: W1212 08:30:36.764697 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0b99c63_9b5a_4ffe_92f4_5fef96b3df72.slice/crio-e2bbe1d935d2ee8f70b92c27fa19611dc420972d7b9b16bbf10d9a297e70cb54 WatchSource:0}: Error finding container e2bbe1d935d2ee8f70b92c27fa19611dc420972d7b9b16bbf10d9a297e70cb54: Status 404 returned error can't find the container with id e2bbe1d935d2ee8f70b92c27fa19611dc420972d7b9b16bbf10d9a297e70cb54 Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.767959 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.860309 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f762879-f3a1-47e2-a6fc-a60dc14701e4" path="/var/lib/kubelet/pods/5f762879-f3a1-47e2-a6fc-a60dc14701e4/volumes" Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.861023 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72","Type":"ContainerStarted","Data":"e2bbe1d935d2ee8f70b92c27fa19611dc420972d7b9b16bbf10d9a297e70cb54"} Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.863322 4867 generic.go:334] "Generic (PLEG): container finished" podID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerID="d2f224e36e1e7f2a13438e91f11e75f05431e99c99e016f55fb3444f6fe12eb8" exitCode=137 Dec 12 08:30:36 crc kubenswrapper[4867]: I1212 08:30:36.863369 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerDied","Data":"d2f224e36e1e7f2a13438e91f11e75f05431e99c99e016f55fb3444f6fe12eb8"} Dec 12 08:30:36 crc kubenswrapper[4867]: E1212 08:30:36.921804 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.070862 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247237 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247604 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gnx8\" (UniqueName: \"kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247644 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247743 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247764 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247807 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247884 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247912 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.247992 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle\") pod \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\" (UID: \"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c\") " Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.248576 4867 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.248589 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.252846 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts" (OuterVolumeSpecName: "scripts") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.255111 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8" (OuterVolumeSpecName: "kube-api-access-8gnx8") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "kube-api-access-8gnx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.282097 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.310460 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.334291 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350666 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350701 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gnx8\" (UniqueName: \"kubernetes.io/projected/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-kube-api-access-8gnx8\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350711 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350719 4867 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350727 4867 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.350734 4867 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-scripts\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.359055 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data" (OuterVolumeSpecName: "config-data") pod "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" (UID: "2a0693b8-b55e-42a7-bfdb-8d3778b89f7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.452793 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.873839 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72","Type":"ContainerStarted","Data":"0aa3a22f9b9014521ff0dc9420330901c59a37782cb5cd2c7df5ffa5be1fd042"} Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.874183 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72","Type":"ContainerStarted","Data":"6e276926fb963cb08de898b93375d5b8842824aa46ee5052be5f80aa0f0e8cd1"} Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.877060 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2a0693b8-b55e-42a7-bfdb-8d3778b89f7c","Type":"ContainerDied","Data":"b981cee93f8d7a795bb68277d9dc83261009b90682783a1cef56bc3a0a731542"} Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.877101 4867 scope.go:117] "RemoveContainer" containerID="d2f224e36e1e7f2a13438e91f11e75f05431e99c99e016f55fb3444f6fe12eb8" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.877130 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.903073 4867 scope.go:117] "RemoveContainer" containerID="2521d068d6818a80a12fd4e6471a955013d8fc1c55c236c60dedb1b10cd46224" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.920290 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.929315 4867 scope.go:117] "RemoveContainer" containerID="46d0fc9d73df7022831d3764ec51c4b2cc6546b4613cb379f789bc779c559340" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.944128 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.960310 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:37 crc kubenswrapper[4867]: E1212 08:30:37.960818 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-notification-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.960842 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-notification-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: E1212 08:30:37.960854 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="sg-core" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.960861 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="sg-core" Dec 12 08:30:37 crc kubenswrapper[4867]: E1212 08:30:37.960876 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="proxy-httpd" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.960886 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="proxy-httpd" Dec 12 08:30:37 crc kubenswrapper[4867]: E1212 08:30:37.960905 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-central-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.960911 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-central-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.961101 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-central-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.961116 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="sg-core" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.961128 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="ceilometer-notification-agent" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.961142 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" containerName="proxy-httpd" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.963454 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.967950 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.968844 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.972651 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.983759 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:37 crc kubenswrapper[4867]: I1212 08:30:37.989316 4867 scope.go:117] "RemoveContainer" containerID="3a1c3b79dd633c2960ba54a2dd6b22e05d0640a8c30ff9353c5b108cb40e936a" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.065932 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-run-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066000 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066357 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-scripts\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066479 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-config-data\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066523 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bls72\" (UniqueName: \"kubernetes.io/projected/50dd7d57-3939-45db-bb38-9e6198ba59be-kube-api-access-bls72\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066619 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066653 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.066737 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-log-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169038 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-run-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169102 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169170 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-scripts\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169271 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-config-data\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169313 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bls72\" (UniqueName: \"kubernetes.io/projected/50dd7d57-3939-45db-bb38-9e6198ba59be-kube-api-access-bls72\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169391 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169418 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169494 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-log-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169947 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-log-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.169953 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/50dd7d57-3939-45db-bb38-9e6198ba59be-run-httpd\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.174679 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-scripts\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.177910 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.179197 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.179539 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-config-data\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.183650 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50dd7d57-3939-45db-bb38-9e6198ba59be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.187591 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bls72\" (UniqueName: \"kubernetes.io/projected/50dd7d57-3939-45db-bb38-9e6198ba59be-kube-api-access-bls72\") pod \"ceilometer-0\" (UID: \"50dd7d57-3939-45db-bb38-9e6198ba59be\") " pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.290960 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.756603 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.855396 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a0693b8-b55e-42a7-bfdb-8d3778b89f7c" path="/var/lib/kubelet/pods/2a0693b8-b55e-42a7-bfdb-8d3778b89f7c/volumes" Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.889277 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72","Type":"ContainerStarted","Data":"a117abe85c3ce7db512016d7e2257f5dfe5098700b156e757a7f82f05f634b72"} Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.890861 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d0b99c63-9b5a-4ffe-92f4-5fef96b3df72","Type":"ContainerStarted","Data":"51b353f2e9049ee9d1b2bb9ea46423d650fbee2d74c5225d5e6c657d9f24560c"} Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.891870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50dd7d57-3939-45db-bb38-9e6198ba59be","Type":"ContainerStarted","Data":"8f33b501b9486d1651b7329dd64e475fd531f5a77ce0b65bad0dc5308f643332"} Dec 12 08:30:38 crc kubenswrapper[4867]: I1212 08:30:38.917821 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.555207175 podStartE2EDuration="3.917799788s" podCreationTimestamp="2025-12-12 08:30:35 +0000 UTC" firstStartedPulling="2025-12-12 08:30:36.767753966 +0000 UTC m=+6124.339135225" lastFinishedPulling="2025-12-12 08:30:38.130346569 +0000 UTC m=+6125.701727838" observedRunningTime="2025-12-12 08:30:38.90860957 +0000 UTC m=+6126.479990829" watchObservedRunningTime="2025-12-12 08:30:38.917799788 +0000 UTC m=+6126.489181057" Dec 12 08:30:39 crc kubenswrapper[4867]: I1212 08:30:39.842003 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:30:39 crc kubenswrapper[4867]: E1212 08:30:39.842600 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:30:39 crc kubenswrapper[4867]: I1212 08:30:39.904383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50dd7d57-3939-45db-bb38-9e6198ba59be","Type":"ContainerStarted","Data":"56c52733833c04c6deb3dff1ed548f11b298c23024dd65e7c6b7e764d4217538"} Dec 12 08:30:39 crc kubenswrapper[4867]: I1212 08:30:39.904727 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50dd7d57-3939-45db-bb38-9e6198ba59be","Type":"ContainerStarted","Data":"6bd4dd4649ae2429c1d2c9a7f0fd184eb8c94bebcf43c2ce1c332c70129d71f1"} Dec 12 08:30:40 crc kubenswrapper[4867]: I1212 08:30:40.920477 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50dd7d57-3939-45db-bb38-9e6198ba59be","Type":"ContainerStarted","Data":"e9d8d3f445186cc13bdd7507d10b92bdf4269c621933fff358d1235e0464772c"} Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.176959 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.178964 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.187308 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.210289 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.339177 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.339324 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmdjd\" (UniqueName: \"kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.339420 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.339767 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.339950 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.340012 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.441562 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442014 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442181 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442344 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442587 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442745 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmdjd\" (UniqueName: \"kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.442906 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.443158 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.443293 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.443568 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.444034 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.469257 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmdjd\" (UniqueName: \"kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd\") pod \"dnsmasq-dns-64489754bc-xrmm7\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:41 crc kubenswrapper[4867]: I1212 08:30:41.504332 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.046510 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:30:42 crc kubenswrapper[4867]: W1212 08:30:42.048846 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1c28265_9e79_4cd5_b720_10a24c415113.slice/crio-529f1d020ef70d5e79189d7e288464e66f02b83a800301444fd6131db276f4fb WatchSource:0}: Error finding container 529f1d020ef70d5e79189d7e288464e66f02b83a800301444fd6131db276f4fb: Status 404 returned error can't find the container with id 529f1d020ef70d5e79189d7e288464e66f02b83a800301444fd6131db276f4fb Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.939042 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"50dd7d57-3939-45db-bb38-9e6198ba59be","Type":"ContainerStarted","Data":"e4ce5094f86407f77004084e0139dbae672825f5dda44ece8172f6d8b49d71b0"} Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.939635 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.941099 4867 generic.go:334] "Generic (PLEG): container finished" podID="c1c28265-9e79-4cd5-b720-10a24c415113" containerID="605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d" exitCode=0 Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.941144 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" event={"ID":"c1c28265-9e79-4cd5-b720-10a24c415113","Type":"ContainerDied","Data":"605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d"} Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.941190 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" event={"ID":"c1c28265-9e79-4cd5-b720-10a24c415113","Type":"ContainerStarted","Data":"529f1d020ef70d5e79189d7e288464e66f02b83a800301444fd6131db276f4fb"} Dec 12 08:30:42 crc kubenswrapper[4867]: I1212 08:30:42.958527 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.348462358 podStartE2EDuration="5.958505209s" podCreationTimestamp="2025-12-12 08:30:37 +0000 UTC" firstStartedPulling="2025-12-12 08:30:38.752355855 +0000 UTC m=+6126.323737124" lastFinishedPulling="2025-12-12 08:30:42.362398706 +0000 UTC m=+6129.933779975" observedRunningTime="2025-12-12 08:30:42.957369252 +0000 UTC m=+6130.528750541" watchObservedRunningTime="2025-12-12 08:30:42.958505209 +0000 UTC m=+6130.529886478" Dec 12 08:30:43 crc kubenswrapper[4867]: I1212 08:30:43.952184 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" event={"ID":"c1c28265-9e79-4cd5-b720-10a24c415113","Type":"ContainerStarted","Data":"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713"} Dec 12 08:30:43 crc kubenswrapper[4867]: I1212 08:30:43.953071 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:43 crc kubenswrapper[4867]: I1212 08:30:43.980613 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" podStartSLOduration=2.980594608 podStartE2EDuration="2.980594608s" podCreationTimestamp="2025-12-12 08:30:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:30:43.971874401 +0000 UTC m=+6131.543255670" watchObservedRunningTime="2025-12-12 08:30:43.980594608 +0000 UTC m=+6131.551975877" Dec 12 08:30:47 crc kubenswrapper[4867]: E1212 08:30:47.195517 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.042475 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ksqq5"] Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.052833 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4768-account-create-update-rcgbk"] Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.063240 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ksqq5"] Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.071249 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4768-account-create-update-rcgbk"] Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.850056 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22edfff0-ed98-4a5d-b74b-88c8bc029cb7" path="/var/lib/kubelet/pods/22edfff0-ed98-4a5d-b74b-88c8bc029cb7/volumes" Dec 12 08:30:48 crc kubenswrapper[4867]: I1212 08:30:48.850655 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ef59d9-3ad2-4652-9ed3-69f1798c89d0" path="/var/lib/kubelet/pods/65ef59d9-3ad2-4652-9ed3-69f1798c89d0/volumes" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.506261 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.577660 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.577949 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="dnsmasq-dns" containerID="cri-o://a5d29a5ea3cfb4fc1f1d752ab310eaaf6d1a1c51b97aafb0d88df751920034c0" gracePeriod=10 Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.791614 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69b4d64989-z2k6r"] Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.799622 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.814509 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4d64989-z2k6r"] Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.917568 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.917712 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9fkj\" (UniqueName: \"kubernetes.io/projected/671819cf-f023-49bf-9e27-0a3e0c1ec578-kube-api-access-s9fkj\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.917738 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-openstack-cell1\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.917768 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-dns-svc\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.917939 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:51 crc kubenswrapper[4867]: I1212 08:30:51.918024 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-config\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021408 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021563 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9fkj\" (UniqueName: \"kubernetes.io/projected/671819cf-f023-49bf-9e27-0a3e0c1ec578-kube-api-access-s9fkj\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021594 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-openstack-cell1\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021626 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-dns-svc\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021741 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.021813 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-config\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.023267 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-config\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.023990 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-openstack-cell1\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.025813 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-dns-svc\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.025187 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.026356 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/671819cf-f023-49bf-9e27-0a3e0c1ec578-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.051906 4867 generic.go:334] "Generic (PLEG): container finished" podID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerID="a5d29a5ea3cfb4fc1f1d752ab310eaaf6d1a1c51b97aafb0d88df751920034c0" exitCode=0 Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.052002 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" event={"ID":"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9","Type":"ContainerDied","Data":"a5d29a5ea3cfb4fc1f1d752ab310eaaf6d1a1c51b97aafb0d88df751920034c0"} Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.060129 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9fkj\" (UniqueName: \"kubernetes.io/projected/671819cf-f023-49bf-9e27-0a3e0c1ec578-kube-api-access-s9fkj\") pod \"dnsmasq-dns-69b4d64989-z2k6r\" (UID: \"671819cf-f023-49bf-9e27-0a3e0c1ec578\") " pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.145919 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.171435 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.226494 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config\") pod \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.227130 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qffx\" (UniqueName: \"kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx\") pod \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.227169 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc\") pod \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.227244 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb\") pod \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.227274 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb\") pod \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\" (UID: \"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9\") " Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.237436 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx" (OuterVolumeSpecName: "kube-api-access-4qffx") pod "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" (UID: "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9"). InnerVolumeSpecName "kube-api-access-4qffx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.313991 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" (UID: "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.329584 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qffx\" (UniqueName: \"kubernetes.io/projected/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-kube-api-access-4qffx\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.329624 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.342738 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config" (OuterVolumeSpecName: "config") pod "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" (UID: "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.342846 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" (UID: "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.343420 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" (UID: "67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.439532 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.440552 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.440594 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:30:52 crc kubenswrapper[4867]: I1212 08:30:52.666346 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4d64989-z2k6r"] Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.067250 4867 generic.go:334] "Generic (PLEG): container finished" podID="671819cf-f023-49bf-9e27-0a3e0c1ec578" containerID="59415a8bfd3c4b2488a672cb5ff7adaa2a5d971494d8cd62944806757f7bd2d3" exitCode=0 Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.067355 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" event={"ID":"671819cf-f023-49bf-9e27-0a3e0c1ec578","Type":"ContainerDied","Data":"59415a8bfd3c4b2488a672cb5ff7adaa2a5d971494d8cd62944806757f7bd2d3"} Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.067899 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" event={"ID":"671819cf-f023-49bf-9e27-0a3e0c1ec578","Type":"ContainerStarted","Data":"0fbed4164634c03cf806a0095f6e4b56706c049e1e29d896763e588ac0d6b5ca"} Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.070460 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" event={"ID":"67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9","Type":"ContainerDied","Data":"923130bb40ad8626bbb78e02c3b37c6afdc0602d1b514d98ddbbfcb9ebbd3563"} Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.070554 4867 scope.go:117] "RemoveContainer" containerID="a5d29a5ea3cfb4fc1f1d752ab310eaaf6d1a1c51b97aafb0d88df751920034c0" Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.070739 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd7f7d77f-dm22f" Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.124504 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.143607 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bd7f7d77f-dm22f"] Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.268197 4867 scope.go:117] "RemoveContainer" containerID="072c8049ad88ff4bf15d8cc3db9a8a70266ebaf79ffd281fa871d1b740e79d98" Dec 12 08:30:53 crc kubenswrapper[4867]: I1212 08:30:53.838356 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:30:53 crc kubenswrapper[4867]: E1212 08:30:53.839105 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:30:54 crc kubenswrapper[4867]: I1212 08:30:54.081322 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" event={"ID":"671819cf-f023-49bf-9e27-0a3e0c1ec578","Type":"ContainerStarted","Data":"44247b7867ab2300d4b66062232cfe10435365c37cc634b23e5fdcf5689fccde"} Dec 12 08:30:54 crc kubenswrapper[4867]: I1212 08:30:54.082716 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:30:54 crc kubenswrapper[4867]: I1212 08:30:54.115607 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" podStartSLOduration=3.115580991 podStartE2EDuration="3.115580991s" podCreationTimestamp="2025-12-12 08:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 08:30:54.106756452 +0000 UTC m=+6141.678137721" watchObservedRunningTime="2025-12-12 08:30:54.115580991 +0000 UTC m=+6141.686962260" Dec 12 08:30:54 crc kubenswrapper[4867]: I1212 08:30:54.850852 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" path="/var/lib/kubelet/pods/67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9/volumes" Dec 12 08:30:57 crc kubenswrapper[4867]: E1212 08:30:57.469087 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:31:00 crc kubenswrapper[4867]: I1212 08:31:00.147383 4867 scope.go:117] "RemoveContainer" containerID="a6ee51b17db7d5155140ba45a07c9a60837aa49608c73b8eb7204e2444fd7a64" Dec 12 08:31:00 crc kubenswrapper[4867]: I1212 08:31:00.176662 4867 scope.go:117] "RemoveContainer" containerID="2101f0952f228d0909a976df96b4c591d3c99076d3390137125da4e5a04e04a1" Dec 12 08:31:00 crc kubenswrapper[4867]: I1212 08:31:00.241935 4867 scope.go:117] "RemoveContainer" containerID="ab8dd21f54dce96713b525c86ade91bc53a534754bdb926c40507f9a80a0f5d5" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.181348 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69b4d64989-z2k6r" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.357607 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.358213 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="dnsmasq-dns" containerID="cri-o://6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713" gracePeriod=10 Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.880861 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912691 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912749 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912794 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmdjd\" (UniqueName: \"kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912845 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912881 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.912963 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb\") pod \"c1c28265-9e79-4cd5-b720-10a24c415113\" (UID: \"c1c28265-9e79-4cd5-b720-10a24c415113\") " Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.918760 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd" (OuterVolumeSpecName: "kube-api-access-hmdjd") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "kube-api-access-hmdjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.965273 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config" (OuterVolumeSpecName: "config") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.967558 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.968836 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.982512 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:31:02 crc kubenswrapper[4867]: I1212 08:31:02.995120 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c1c28265-9e79-4cd5-b720-10a24c415113" (UID: "c1c28265-9e79-4cd5-b720-10a24c415113"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.017525 4867 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-config\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.017926 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.018023 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmdjd\" (UniqueName: \"kubernetes.io/projected/c1c28265-9e79-4cd5-b720-10a24c415113-kube-api-access-hmdjd\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.018091 4867 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.018142 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.018197 4867 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1c28265-9e79-4cd5-b720-10a24c415113-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.185416 4867 generic.go:334] "Generic (PLEG): container finished" podID="c1c28265-9e79-4cd5-b720-10a24c415113" containerID="6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713" exitCode=0 Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.185468 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.185473 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" event={"ID":"c1c28265-9e79-4cd5-b720-10a24c415113","Type":"ContainerDied","Data":"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713"} Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.185516 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64489754bc-xrmm7" event={"ID":"c1c28265-9e79-4cd5-b720-10a24c415113","Type":"ContainerDied","Data":"529f1d020ef70d5e79189d7e288464e66f02b83a800301444fd6131db276f4fb"} Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.185537 4867 scope.go:117] "RemoveContainer" containerID="6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.214572 4867 scope.go:117] "RemoveContainer" containerID="605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.230881 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.239760 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64489754bc-xrmm7"] Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.252107 4867 scope.go:117] "RemoveContainer" containerID="6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713" Dec 12 08:31:03 crc kubenswrapper[4867]: E1212 08:31:03.252897 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713\": container with ID starting with 6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713 not found: ID does not exist" containerID="6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.253565 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713"} err="failed to get container status \"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713\": rpc error: code = NotFound desc = could not find container \"6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713\": container with ID starting with 6bf142d890a5cb2d2725dc45738e5d929851114a3f4bb1dd613416294ebe9713 not found: ID does not exist" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.253732 4867 scope.go:117] "RemoveContainer" containerID="605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d" Dec 12 08:31:03 crc kubenswrapper[4867]: E1212 08:31:03.256659 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d\": container with ID starting with 605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d not found: ID does not exist" containerID="605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d" Dec 12 08:31:03 crc kubenswrapper[4867]: I1212 08:31:03.256700 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d"} err="failed to get container status \"605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d\": rpc error: code = NotFound desc = could not find container \"605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d\": container with ID starting with 605115e0294dfe0543aeb4107a968ed7e21f9e4eb312113eb6f32f1002b62d2d not found: ID does not exist" Dec 12 08:31:04 crc kubenswrapper[4867]: I1212 08:31:04.850950 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" path="/var/lib/kubelet/pods/c1c28265-9e79-4cd5-b720-10a24c415113/volumes" Dec 12 08:31:05 crc kubenswrapper[4867]: I1212 08:31:05.838991 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:31:05 crc kubenswrapper[4867]: E1212 08:31:05.839201 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:31:07 crc kubenswrapper[4867]: E1212 08:31:07.749344 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:31:08 crc kubenswrapper[4867]: I1212 08:31:08.299758 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.602359 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk"] Dec 12 08:31:12 crc kubenswrapper[4867]: E1212 08:31:12.603049 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603061 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: E1212 08:31:12.603074 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="init" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603079 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="init" Dec 12 08:31:12 crc kubenswrapper[4867]: E1212 08:31:12.603106 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="init" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603112 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="init" Dec 12 08:31:12 crc kubenswrapper[4867]: E1212 08:31:12.603136 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603141 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603356 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1c28265-9e79-4cd5-b720-10a24c415113" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.603371 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="67bb9d1e-b7f9-46f7-a5ce-2a9e17e6c4d9" containerName="dnsmasq-dns" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.604072 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.606178 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.606181 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.606243 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.606522 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.625011 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk"] Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.805656 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.805746 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.805878 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.805917 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw8sl\" (UniqueName: \"kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.908262 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.908373 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.908492 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.908534 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw8sl\" (UniqueName: \"kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.914337 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.914525 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.926080 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:12 crc kubenswrapper[4867]: I1212 08:31:12.928931 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw8sl\" (UniqueName: \"kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:13 crc kubenswrapper[4867]: I1212 08:31:13.042028 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-457vz"] Dec 12 08:31:13 crc kubenswrapper[4867]: I1212 08:31:13.053935 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-457vz"] Dec 12 08:31:13 crc kubenswrapper[4867]: I1212 08:31:13.222139 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:13 crc kubenswrapper[4867]: I1212 08:31:13.792110 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk"] Dec 12 08:31:14 crc kubenswrapper[4867]: I1212 08:31:14.308809 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" event={"ID":"821d9b97-2dd4-456d-8dbc-5bb0da1685f7","Type":"ContainerStarted","Data":"9f6fd2fa7525cb8715164de593eb5854e5242d53a45791fcc68a90608f406366"} Dec 12 08:31:14 crc kubenswrapper[4867]: I1212 08:31:14.854841 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fceb31cc-12c6-46ab-85d2-107a5b9696d3" path="/var/lib/kubelet/pods/fceb31cc-12c6-46ab-85d2-107a5b9696d3/volumes" Dec 12 08:31:18 crc kubenswrapper[4867]: E1212 08:31:18.017413 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:31:19 crc kubenswrapper[4867]: I1212 08:31:19.838777 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:31:19 crc kubenswrapper[4867]: E1212 08:31:19.839540 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:31:22 crc kubenswrapper[4867]: I1212 08:31:22.390383 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" event={"ID":"821d9b97-2dd4-456d-8dbc-5bb0da1685f7","Type":"ContainerStarted","Data":"5579b7cfb3755ae1fd86a163246e0037ec9bcf35280173c3dcdb9abe3f1c84ce"} Dec 12 08:31:22 crc kubenswrapper[4867]: I1212 08:31:22.417382 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" podStartSLOduration=2.243111483 podStartE2EDuration="10.417356699s" podCreationTimestamp="2025-12-12 08:31:12 +0000 UTC" firstStartedPulling="2025-12-12 08:31:13.793577495 +0000 UTC m=+6161.364958754" lastFinishedPulling="2025-12-12 08:31:21.967822701 +0000 UTC m=+6169.539203970" observedRunningTime="2025-12-12 08:31:22.412757755 +0000 UTC m=+6169.984139034" watchObservedRunningTime="2025-12-12 08:31:22.417356699 +0000 UTC m=+6169.988737968" Dec 12 08:31:28 crc kubenswrapper[4867]: E1212 08:31:28.252436 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f762879_f3a1_47e2_a6fc_a60dc14701e4.slice/crio-f1f5abebc98b67213e9596fce7214a6478e98c037254cd27a28e4646f0e51a1c\": RecentStats: unable to find data in memory cache]" Dec 12 08:31:31 crc kubenswrapper[4867]: I1212 08:31:31.837843 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:31:31 crc kubenswrapper[4867]: E1212 08:31:31.838705 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:31:34 crc kubenswrapper[4867]: I1212 08:31:34.512131 4867 generic.go:334] "Generic (PLEG): container finished" podID="821d9b97-2dd4-456d-8dbc-5bb0da1685f7" containerID="5579b7cfb3755ae1fd86a163246e0037ec9bcf35280173c3dcdb9abe3f1c84ce" exitCode=0 Dec 12 08:31:34 crc kubenswrapper[4867]: I1212 08:31:34.512234 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" event={"ID":"821d9b97-2dd4-456d-8dbc-5bb0da1685f7","Type":"ContainerDied","Data":"5579b7cfb3755ae1fd86a163246e0037ec9bcf35280173c3dcdb9abe3f1c84ce"} Dec 12 08:31:35 crc kubenswrapper[4867]: I1212 08:31:35.932162 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:35 crc kubenswrapper[4867]: I1212 08:31:35.991515 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key\") pod \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " Dec 12 08:31:35 crc kubenswrapper[4867]: I1212 08:31:35.991622 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw8sl\" (UniqueName: \"kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl\") pod \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " Dec 12 08:31:35 crc kubenswrapper[4867]: I1212 08:31:35.992339 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle\") pod \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " Dec 12 08:31:35 crc kubenswrapper[4867]: I1212 08:31:35.992611 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory\") pod \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\" (UID: \"821d9b97-2dd4-456d-8dbc-5bb0da1685f7\") " Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.011721 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl" (OuterVolumeSpecName: "kube-api-access-mw8sl") pod "821d9b97-2dd4-456d-8dbc-5bb0da1685f7" (UID: "821d9b97-2dd4-456d-8dbc-5bb0da1685f7"). InnerVolumeSpecName "kube-api-access-mw8sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.021376 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "821d9b97-2dd4-456d-8dbc-5bb0da1685f7" (UID: "821d9b97-2dd4-456d-8dbc-5bb0da1685f7"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.102627 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw8sl\" (UniqueName: \"kubernetes.io/projected/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-kube-api-access-mw8sl\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.102661 4867 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.109390 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory" (OuterVolumeSpecName: "inventory") pod "821d9b97-2dd4-456d-8dbc-5bb0da1685f7" (UID: "821d9b97-2dd4-456d-8dbc-5bb0da1685f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.117667 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "821d9b97-2dd4-456d-8dbc-5bb0da1685f7" (UID: "821d9b97-2dd4-456d-8dbc-5bb0da1685f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.204138 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.204174 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/821d9b97-2dd4-456d-8dbc-5bb0da1685f7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.535656 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" event={"ID":"821d9b97-2dd4-456d-8dbc-5bb0da1685f7","Type":"ContainerDied","Data":"9f6fd2fa7525cb8715164de593eb5854e5242d53a45791fcc68a90608f406366"} Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.535694 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f6fd2fa7525cb8715164de593eb5854e5242d53a45791fcc68a90608f406366" Dec 12 08:31:36 crc kubenswrapper[4867]: I1212 08:31:36.535723 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk" Dec 12 08:31:42 crc kubenswrapper[4867]: I1212 08:31:42.086162 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-rbk8h"] Dec 12 08:31:42 crc kubenswrapper[4867]: I1212 08:31:42.110751 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-rbk8h"] Dec 12 08:31:42 crc kubenswrapper[4867]: I1212 08:31:42.846389 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:31:42 crc kubenswrapper[4867]: E1212 08:31:42.846905 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:31:42 crc kubenswrapper[4867]: I1212 08:31:42.848881 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97a1ec6-d7fc-4f2d-a063-891aee80a17b" path="/var/lib/kubelet/pods/f97a1ec6-d7fc-4f2d-a063-891aee80a17b/volumes" Dec 12 08:31:43 crc kubenswrapper[4867]: I1212 08:31:43.028863 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f55b-account-create-update-db6tz"] Dec 12 08:31:43 crc kubenswrapper[4867]: I1212 08:31:43.040867 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f55b-account-create-update-db6tz"] Dec 12 08:31:44 crc kubenswrapper[4867]: I1212 08:31:44.859358 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a1ef07e-6698-40ab-9251-ca02f2e2726f" path="/var/lib/kubelet/pods/9a1ef07e-6698-40ab-9251-ca02f2e2726f/volumes" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.376928 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn"] Dec 12 08:31:45 crc kubenswrapper[4867]: E1212 08:31:45.379356 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821d9b97-2dd4-456d-8dbc-5bb0da1685f7" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.379380 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="821d9b97-2dd4-456d-8dbc-5bb0da1685f7" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.379589 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="821d9b97-2dd4-456d-8dbc-5bb0da1685f7" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.380326 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.382931 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.385093 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.385259 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.385703 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.391395 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn"] Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.440273 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.440353 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.440665 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.440744 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.542684 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.542773 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.542798 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.542932 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.548823 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.548865 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.549459 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.559302 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:45 crc kubenswrapper[4867]: I1212 08:31:45.778283 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:31:46 crc kubenswrapper[4867]: I1212 08:31:46.325049 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn"] Dec 12 08:31:46 crc kubenswrapper[4867]: I1212 08:31:46.646394 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" event={"ID":"55c56e88-ce42-4abb-bdfb-e68fd154fea6","Type":"ContainerStarted","Data":"dcbed3c754a7675577e6ed09ef6f17ea1911e4905f19f5a21f5c563f1efbcc3b"} Dec 12 08:31:47 crc kubenswrapper[4867]: I1212 08:31:47.656916 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" event={"ID":"55c56e88-ce42-4abb-bdfb-e68fd154fea6","Type":"ContainerStarted","Data":"24d2676e5cc5576ee28d4d9fb8b71a1e290498b3cd6d2a9cdd701926d35228c8"} Dec 12 08:31:47 crc kubenswrapper[4867]: I1212 08:31:47.673162 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" podStartSLOduration=1.859736483 podStartE2EDuration="2.673140796s" podCreationTimestamp="2025-12-12 08:31:45 +0000 UTC" firstStartedPulling="2025-12-12 08:31:46.336346303 +0000 UTC m=+6193.907727572" lastFinishedPulling="2025-12-12 08:31:47.149750616 +0000 UTC m=+6194.721131885" observedRunningTime="2025-12-12 08:31:47.671130236 +0000 UTC m=+6195.242511515" watchObservedRunningTime="2025-12-12 08:31:47.673140796 +0000 UTC m=+6195.244522065" Dec 12 08:31:52 crc kubenswrapper[4867]: I1212 08:31:52.031962 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zzx5d"] Dec 12 08:31:52 crc kubenswrapper[4867]: I1212 08:31:52.043971 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zzx5d"] Dec 12 08:31:52 crc kubenswrapper[4867]: I1212 08:31:52.859379 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e38feb-db94-479d-9856-cd92df3741e1" path="/var/lib/kubelet/pods/62e38feb-db94-479d-9856-cd92df3741e1/volumes" Dec 12 08:31:53 crc kubenswrapper[4867]: I1212 08:31:53.838833 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:31:53 crc kubenswrapper[4867]: E1212 08:31:53.839768 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.433474 4867 scope.go:117] "RemoveContainer" containerID="9f45fee95e5683aa30b40dcb21bd9c869e772b0b70aeb11f1072cef1ee6c87c6" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.465491 4867 scope.go:117] "RemoveContainer" containerID="ffdd0a96d087e818753d60df3de929e64e6477dcef0a15cb4b5d224a5762569c" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.489204 4867 scope.go:117] "RemoveContainer" containerID="3582b78b6abced80afd095968cb3ce17490aefa756cb1c6f43443f966674e197" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.564310 4867 scope.go:117] "RemoveContainer" containerID="ea428821739f255818a70b3c565d6503762056608cab40f46c73e1f01bafd3fd" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.576546 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.579747 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.609637 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.624813 4867 scope.go:117] "RemoveContainer" containerID="c45c539948831f6a2c7314b9f5a259c1250df1bb745656746800276c7aacd025" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.668071 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.668179 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.668400 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-822zt\" (UniqueName: \"kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.769924 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-822zt\" (UniqueName: \"kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.769986 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.770040 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.770874 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.770912 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.791634 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-822zt\" (UniqueName: \"kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt\") pod \"redhat-marketplace-44xlq\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.825845 4867 scope.go:117] "RemoveContainer" containerID="41cd9fb6589c4a22a975a78cd6c202ebd083af10a78d50917a3bc11e69b1b1d0" Dec 12 08:32:00 crc kubenswrapper[4867]: I1212 08:32:00.968006 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:01 crc kubenswrapper[4867]: I1212 08:32:01.413203 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:01 crc kubenswrapper[4867]: I1212 08:32:01.796482 4867 generic.go:334] "Generic (PLEG): container finished" podID="0bd940ce-557f-4502-945c-6c44d988dabf" containerID="b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0" exitCode=0 Dec 12 08:32:01 crc kubenswrapper[4867]: I1212 08:32:01.796548 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerDied","Data":"b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0"} Dec 12 08:32:01 crc kubenswrapper[4867]: I1212 08:32:01.796572 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerStarted","Data":"704a6461d129ee6f5fa5b0cf8fb2c118889dc414cdf4f94ea03667631afbf96b"} Dec 12 08:32:02 crc kubenswrapper[4867]: I1212 08:32:02.810040 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerStarted","Data":"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36"} Dec 12 08:32:03 crc kubenswrapper[4867]: I1212 08:32:03.822801 4867 generic.go:334] "Generic (PLEG): container finished" podID="0bd940ce-557f-4502-945c-6c44d988dabf" containerID="b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36" exitCode=0 Dec 12 08:32:03 crc kubenswrapper[4867]: I1212 08:32:03.822896 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerDied","Data":"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36"} Dec 12 08:32:05 crc kubenswrapper[4867]: I1212 08:32:05.844488 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerStarted","Data":"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b"} Dec 12 08:32:05 crc kubenswrapper[4867]: I1212 08:32:05.868372 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-44xlq" podStartSLOduration=2.988183987 podStartE2EDuration="5.868281675s" podCreationTimestamp="2025-12-12 08:32:00 +0000 UTC" firstStartedPulling="2025-12-12 08:32:01.80164629 +0000 UTC m=+6209.373027559" lastFinishedPulling="2025-12-12 08:32:04.681743978 +0000 UTC m=+6212.253125247" observedRunningTime="2025-12-12 08:32:05.864398089 +0000 UTC m=+6213.435779358" watchObservedRunningTime="2025-12-12 08:32:05.868281675 +0000 UTC m=+6213.439662944" Dec 12 08:32:06 crc kubenswrapper[4867]: I1212 08:32:06.838865 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:32:06 crc kubenswrapper[4867]: E1212 08:32:06.839194 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:32:10 crc kubenswrapper[4867]: I1212 08:32:10.969074 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:10 crc kubenswrapper[4867]: I1212 08:32:10.969930 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:11 crc kubenswrapper[4867]: I1212 08:32:11.033292 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:11 crc kubenswrapper[4867]: I1212 08:32:11.952165 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:11 crc kubenswrapper[4867]: I1212 08:32:11.997772 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:13 crc kubenswrapper[4867]: I1212 08:32:13.915421 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-44xlq" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="registry-server" containerID="cri-o://744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b" gracePeriod=2 Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.372608 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.465396 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-822zt\" (UniqueName: \"kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt\") pod \"0bd940ce-557f-4502-945c-6c44d988dabf\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.465492 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities\") pod \"0bd940ce-557f-4502-945c-6c44d988dabf\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.465581 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content\") pod \"0bd940ce-557f-4502-945c-6c44d988dabf\" (UID: \"0bd940ce-557f-4502-945c-6c44d988dabf\") " Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.466823 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities" (OuterVolumeSpecName: "utilities") pod "0bd940ce-557f-4502-945c-6c44d988dabf" (UID: "0bd940ce-557f-4502-945c-6c44d988dabf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.470718 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt" (OuterVolumeSpecName: "kube-api-access-822zt") pod "0bd940ce-557f-4502-945c-6c44d988dabf" (UID: "0bd940ce-557f-4502-945c-6c44d988dabf"). InnerVolumeSpecName "kube-api-access-822zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.485894 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bd940ce-557f-4502-945c-6c44d988dabf" (UID: "0bd940ce-557f-4502-945c-6c44d988dabf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.567822 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-822zt\" (UniqueName: \"kubernetes.io/projected/0bd940ce-557f-4502-945c-6c44d988dabf-kube-api-access-822zt\") on node \"crc\" DevicePath \"\"" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.567940 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.567968 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd940ce-557f-4502-945c-6c44d988dabf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.925983 4867 generic.go:334] "Generic (PLEG): container finished" podID="0bd940ce-557f-4502-945c-6c44d988dabf" containerID="744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b" exitCode=0 Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.926070 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerDied","Data":"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b"} Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.926353 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44xlq" event={"ID":"0bd940ce-557f-4502-945c-6c44d988dabf","Type":"ContainerDied","Data":"704a6461d129ee6f5fa5b0cf8fb2c118889dc414cdf4f94ea03667631afbf96b"} Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.926373 4867 scope.go:117] "RemoveContainer" containerID="744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.926086 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44xlq" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.959866 4867 scope.go:117] "RemoveContainer" containerID="b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36" Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.960032 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.970878 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-44xlq"] Dec 12 08:32:14 crc kubenswrapper[4867]: I1212 08:32:14.980840 4867 scope.go:117] "RemoveContainer" containerID="b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.035543 4867 scope.go:117] "RemoveContainer" containerID="744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b" Dec 12 08:32:15 crc kubenswrapper[4867]: E1212 08:32:15.036096 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b\": container with ID starting with 744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b not found: ID does not exist" containerID="744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.036268 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b"} err="failed to get container status \"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b\": rpc error: code = NotFound desc = could not find container \"744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b\": container with ID starting with 744ebb0be2d7bf774e38caf064443017bda47134f14d38949047ce3bc9819b4b not found: ID does not exist" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.036304 4867 scope.go:117] "RemoveContainer" containerID="b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36" Dec 12 08:32:15 crc kubenswrapper[4867]: E1212 08:32:15.036627 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36\": container with ID starting with b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36 not found: ID does not exist" containerID="b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.036662 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36"} err="failed to get container status \"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36\": rpc error: code = NotFound desc = could not find container \"b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36\": container with ID starting with b4821c4cf19ee758b893f878fafb76a7d8b86c77eef8c54eebe14a0e5d24fe36 not found: ID does not exist" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.036685 4867 scope.go:117] "RemoveContainer" containerID="b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0" Dec 12 08:32:15 crc kubenswrapper[4867]: E1212 08:32:15.036958 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0\": container with ID starting with b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0 not found: ID does not exist" containerID="b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0" Dec 12 08:32:15 crc kubenswrapper[4867]: I1212 08:32:15.036983 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0"} err="failed to get container status \"b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0\": rpc error: code = NotFound desc = could not find container \"b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0\": container with ID starting with b44a6b6ebbe64e2142ccdec1b3e6d9546d49de8aad8cab166cffaefafb5715f0 not found: ID does not exist" Dec 12 08:32:16 crc kubenswrapper[4867]: I1212 08:32:16.851991 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" path="/var/lib/kubelet/pods/0bd940ce-557f-4502-945c-6c44d988dabf/volumes" Dec 12 08:32:17 crc kubenswrapper[4867]: I1212 08:32:17.838826 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:32:17 crc kubenswrapper[4867]: E1212 08:32:17.839336 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:32:29 crc kubenswrapper[4867]: I1212 08:32:29.838218 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:32:29 crc kubenswrapper[4867]: E1212 08:32:29.838994 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:32:42 crc kubenswrapper[4867]: I1212 08:32:42.844994 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:32:42 crc kubenswrapper[4867]: E1212 08:32:42.846134 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.046592 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-sgljh"] Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.056638 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-r6b5p"] Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.066669 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-r6b5p"] Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.074602 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-sgljh"] Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.850403 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565d5803-c324-4497-a928-a84f18aaea1b" path="/var/lib/kubelet/pods/565d5803-c324-4497-a928-a84f18aaea1b/volumes" Dec 12 08:32:50 crc kubenswrapper[4867]: I1212 08:32:50.851774 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6cd035-6852-4418-ac56-33fa1f5c2a45" path="/var/lib/kubelet/pods/fa6cd035-6852-4418-ac56-33fa1f5c2a45/volumes" Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.039808 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-skvtb"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.053387 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8bcb-account-create-update-cf9b2"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.065262 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5657-account-create-update-tl2gr"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.073978 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0e7e-account-create-update-g6wnx"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.082444 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-skvtb"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.089948 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5657-account-create-update-tl2gr"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.098813 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8bcb-account-create-update-cf9b2"] Dec 12 08:32:51 crc kubenswrapper[4867]: I1212 08:32:51.106755 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0e7e-account-create-update-g6wnx"] Dec 12 08:32:53 crc kubenswrapper[4867]: I1212 08:32:53.007585 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f689bc-82d9-4867-8505-8ccc6eb237ce" path="/var/lib/kubelet/pods/a7f689bc-82d9-4867-8505-8ccc6eb237ce/volumes" Dec 12 08:32:53 crc kubenswrapper[4867]: I1212 08:32:53.009055 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3d2425-2baf-4842-ac92-256766803f72" path="/var/lib/kubelet/pods/ab3d2425-2baf-4842-ac92-256766803f72/volumes" Dec 12 08:32:53 crc kubenswrapper[4867]: I1212 08:32:53.010064 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c59d7f39-1709-49ba-8623-41f86f89e219" path="/var/lib/kubelet/pods/c59d7f39-1709-49ba-8623-41f86f89e219/volumes" Dec 12 08:32:53 crc kubenswrapper[4867]: I1212 08:32:53.011009 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d48bc89d-ef25-4b10-96f0-be3a93c6eff1" path="/var/lib/kubelet/pods/d48bc89d-ef25-4b10-96f0-be3a93c6eff1/volumes" Dec 12 08:32:53 crc kubenswrapper[4867]: I1212 08:32:53.839082 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:32:53 crc kubenswrapper[4867]: E1212 08:32:53.840062 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:33:00 crc kubenswrapper[4867]: I1212 08:33:00.991882 4867 scope.go:117] "RemoveContainer" containerID="128c9e5af0db226180955e86c009c965e29b78ae1b0a222dd70c5084c5e58fb1" Dec 12 08:33:01 crc kubenswrapper[4867]: I1212 08:33:01.028494 4867 scope.go:117] "RemoveContainer" containerID="2279e76358e2439e53b3b8f373b48ec98fedac9fb7af5fc9a20c1a46bfded688" Dec 12 08:33:01 crc kubenswrapper[4867]: I1212 08:33:01.088788 4867 scope.go:117] "RemoveContainer" containerID="4b6aff1f47fdb4eef73e4e956bac34768aa59938f10924fccea1167da4751cb5" Dec 12 08:33:01 crc kubenswrapper[4867]: I1212 08:33:01.132717 4867 scope.go:117] "RemoveContainer" containerID="0d81a53e3e4ed7888e30655ed6df26c22647349f6dbf6483af939613c7b1782c" Dec 12 08:33:01 crc kubenswrapper[4867]: I1212 08:33:01.194760 4867 scope.go:117] "RemoveContainer" containerID="8e7b7f40702c400378f429ccaed298bb3531c5cb8102bf9b995945297ce0904b" Dec 12 08:33:01 crc kubenswrapper[4867]: I1212 08:33:01.264293 4867 scope.go:117] "RemoveContainer" containerID="0d47260215dbab76368bdcd4eafcd308c2f1399f66909dce69490a6928865fdd" Dec 12 08:33:06 crc kubenswrapper[4867]: I1212 08:33:06.838430 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:33:06 crc kubenswrapper[4867]: E1212 08:33:06.839295 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:33:09 crc kubenswrapper[4867]: I1212 08:33:09.037202 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x2wpg"] Dec 12 08:33:09 crc kubenswrapper[4867]: I1212 08:33:09.045961 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x2wpg"] Dec 12 08:33:10 crc kubenswrapper[4867]: I1212 08:33:10.852173 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba8c7c36-4f4c-499c-9c6c-9eda87455841" path="/var/lib/kubelet/pods/ba8c7c36-4f4c-499c-9c6c-9eda87455841/volumes" Dec 12 08:33:19 crc kubenswrapper[4867]: I1212 08:33:19.838771 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:33:19 crc kubenswrapper[4867]: E1212 08:33:19.839644 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.035066 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4g6xv"] Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.046445 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l749f"] Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.055957 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4g6xv"] Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.065164 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l749f"] Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.849667 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a6d51e8-9cc1-48e6-95d4-afeea668e149" path="/var/lib/kubelet/pods/0a6d51e8-9cc1-48e6-95d4-afeea668e149/volumes" Dec 12 08:33:26 crc kubenswrapper[4867]: I1212 08:33:26.850489 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17206bfe-043b-477a-9c1d-d66156fac47b" path="/var/lib/kubelet/pods/17206bfe-043b-477a-9c1d-d66156fac47b/volumes" Dec 12 08:33:34 crc kubenswrapper[4867]: I1212 08:33:34.838098 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:33:35 crc kubenswrapper[4867]: I1212 08:33:35.407107 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d"} Dec 12 08:34:01 crc kubenswrapper[4867]: I1212 08:34:01.471370 4867 scope.go:117] "RemoveContainer" containerID="f345dc8677ec0295f9398d424dfa621a62c935ec78b631f44f42a88720b24f12" Dec 12 08:34:01 crc kubenswrapper[4867]: I1212 08:34:01.511044 4867 scope.go:117] "RemoveContainer" containerID="ad20c942d6a9d58e9927b8f68be8789211e06512e5690aaa3e25a1e31e624190" Dec 12 08:34:01 crc kubenswrapper[4867]: I1212 08:34:01.559973 4867 scope.go:117] "RemoveContainer" containerID="a6edc791ccee03374cf0bc0cf52daef7c1db7fa42c1ce88e510ec21d26bb2e3f" Dec 12 08:34:13 crc kubenswrapper[4867]: I1212 08:34:13.045277 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-c48nw"] Dec 12 08:34:13 crc kubenswrapper[4867]: I1212 08:34:13.053898 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-c48nw"] Dec 12 08:34:14 crc kubenswrapper[4867]: I1212 08:34:14.849575 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="620fe78b-77e2-4dfa-bae8-a56a029cf5e7" path="/var/lib/kubelet/pods/620fe78b-77e2-4dfa-bae8-a56a029cf5e7/volumes" Dec 12 08:35:01 crc kubenswrapper[4867]: I1212 08:35:01.679921 4867 scope.go:117] "RemoveContainer" containerID="dcf41a6869ed2e886079b80edbb9d4a3f839e14309b6092fa8811d6f95c5e8c0" Dec 12 08:35:01 crc kubenswrapper[4867]: I1212 08:35:01.706391 4867 scope.go:117] "RemoveContainer" containerID="6dd9748445cd52bcb74b1787d737bc9d2468ca63ddaf923d21c0e72c273f6c6f" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.792603 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:15 crc kubenswrapper[4867]: E1212 08:35:15.793681 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="extract-content" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.793696 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="extract-content" Dec 12 08:35:15 crc kubenswrapper[4867]: E1212 08:35:15.793743 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="registry-server" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.793751 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="registry-server" Dec 12 08:35:15 crc kubenswrapper[4867]: E1212 08:35:15.793767 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="extract-utilities" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.793776 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="extract-utilities" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.794093 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bd940ce-557f-4502-945c-6c44d988dabf" containerName="registry-server" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.796752 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.802021 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.845222 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v74kr\" (UniqueName: \"kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.845312 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.845408 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.962542 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v74kr\" (UniqueName: \"kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.962633 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.962711 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.963459 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.963703 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:15 crc kubenswrapper[4867]: I1212 08:35:15.987675 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v74kr\" (UniqueName: \"kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr\") pod \"certified-operators-xxw6v\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:16 crc kubenswrapper[4867]: I1212 08:35:16.121821 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:16 crc kubenswrapper[4867]: I1212 08:35:16.703938 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:17 crc kubenswrapper[4867]: I1212 08:35:17.414841 4867 generic.go:334] "Generic (PLEG): container finished" podID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerID="db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975" exitCode=0 Dec 12 08:35:17 crc kubenswrapper[4867]: I1212 08:35:17.414885 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerDied","Data":"db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975"} Dec 12 08:35:17 crc kubenswrapper[4867]: I1212 08:35:17.414908 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerStarted","Data":"e3aa1c4d2e13b25ef8aca2d0311aa51ee237da6a5fabd5fe062b2bc481540b1c"} Dec 12 08:35:19 crc kubenswrapper[4867]: I1212 08:35:19.439571 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerStarted","Data":"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6"} Dec 12 08:35:20 crc kubenswrapper[4867]: I1212 08:35:20.450005 4867 generic.go:334] "Generic (PLEG): container finished" podID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerID="443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6" exitCode=0 Dec 12 08:35:20 crc kubenswrapper[4867]: I1212 08:35:20.450061 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerDied","Data":"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6"} Dec 12 08:35:21 crc kubenswrapper[4867]: I1212 08:35:21.464014 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerStarted","Data":"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5"} Dec 12 08:35:21 crc kubenswrapper[4867]: I1212 08:35:21.484710 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xxw6v" podStartSLOduration=2.975674202 podStartE2EDuration="6.484693018s" podCreationTimestamp="2025-12-12 08:35:15 +0000 UTC" firstStartedPulling="2025-12-12 08:35:17.419706803 +0000 UTC m=+6404.991088062" lastFinishedPulling="2025-12-12 08:35:20.928725609 +0000 UTC m=+6408.500106878" observedRunningTime="2025-12-12 08:35:21.482184915 +0000 UTC m=+6409.053566204" watchObservedRunningTime="2025-12-12 08:35:21.484693018 +0000 UTC m=+6409.056074287" Dec 12 08:35:26 crc kubenswrapper[4867]: I1212 08:35:26.122408 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:26 crc kubenswrapper[4867]: I1212 08:35:26.123018 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:26 crc kubenswrapper[4867]: I1212 08:35:26.168635 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:26 crc kubenswrapper[4867]: I1212 08:35:26.583423 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:32 crc kubenswrapper[4867]: I1212 08:35:32.754452 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:32 crc kubenswrapper[4867]: I1212 08:35:32.755271 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xxw6v" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="registry-server" containerID="cri-o://2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5" gracePeriod=2 Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.212198 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.222007 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v74kr\" (UniqueName: \"kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr\") pod \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.222207 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities\") pod \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.222342 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content\") pod \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\" (UID: \"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3\") " Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.223446 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities" (OuterVolumeSpecName: "utilities") pod "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" (UID: "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.224011 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.228702 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr" (OuterVolumeSpecName: "kube-api-access-v74kr") pod "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" (UID: "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3"). InnerVolumeSpecName "kube-api-access-v74kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.281182 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" (UID: "bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.326617 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.326674 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v74kr\" (UniqueName: \"kubernetes.io/projected/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3-kube-api-access-v74kr\") on node \"crc\" DevicePath \"\"" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.568659 4867 generic.go:334] "Generic (PLEG): container finished" podID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerID="2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5" exitCode=0 Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.568718 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerDied","Data":"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5"} Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.568749 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxw6v" event={"ID":"bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3","Type":"ContainerDied","Data":"e3aa1c4d2e13b25ef8aca2d0311aa51ee237da6a5fabd5fe062b2bc481540b1c"} Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.568756 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxw6v" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.568767 4867 scope.go:117] "RemoveContainer" containerID="2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.606409 4867 scope.go:117] "RemoveContainer" containerID="443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.615772 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.625042 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xxw6v"] Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.630553 4867 scope.go:117] "RemoveContainer" containerID="db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.676432 4867 scope.go:117] "RemoveContainer" containerID="2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5" Dec 12 08:35:33 crc kubenswrapper[4867]: E1212 08:35:33.676936 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5\": container with ID starting with 2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5 not found: ID does not exist" containerID="2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.676979 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5"} err="failed to get container status \"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5\": rpc error: code = NotFound desc = could not find container \"2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5\": container with ID starting with 2d0b89df9a79d5e9ee76990fc18dd2a91fd58da6376d2e3379a17026fc134fe5 not found: ID does not exist" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.677003 4867 scope.go:117] "RemoveContainer" containerID="443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6" Dec 12 08:35:33 crc kubenswrapper[4867]: E1212 08:35:33.677296 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6\": container with ID starting with 443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6 not found: ID does not exist" containerID="443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.677326 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6"} err="failed to get container status \"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6\": rpc error: code = NotFound desc = could not find container \"443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6\": container with ID starting with 443bf7517679f1d1bb2c5ac5f55e63ac54c4b189c5635d9921830263985ccdf6 not found: ID does not exist" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.677356 4867 scope.go:117] "RemoveContainer" containerID="db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975" Dec 12 08:35:33 crc kubenswrapper[4867]: E1212 08:35:33.677729 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975\": container with ID starting with db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975 not found: ID does not exist" containerID="db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975" Dec 12 08:35:33 crc kubenswrapper[4867]: I1212 08:35:33.677756 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975"} err="failed to get container status \"db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975\": rpc error: code = NotFound desc = could not find container \"db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975\": container with ID starting with db233b9b900ee027d75df959b9c91c4dacbaaeb22e0392e37e2d6ff25820d975 not found: ID does not exist" Dec 12 08:35:34 crc kubenswrapper[4867]: I1212 08:35:34.852803 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" path="/var/lib/kubelet/pods/bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3/volumes" Dec 12 08:35:58 crc kubenswrapper[4867]: I1212 08:35:58.988537 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:35:58 crc kubenswrapper[4867]: I1212 08:35:58.989124 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.924876 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:21 crc kubenswrapper[4867]: E1212 08:36:21.926057 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="extract-content" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.926078 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="extract-content" Dec 12 08:36:21 crc kubenswrapper[4867]: E1212 08:36:21.926102 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="extract-utilities" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.926114 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="extract-utilities" Dec 12 08:36:21 crc kubenswrapper[4867]: E1212 08:36:21.926156 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="registry-server" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.926164 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="registry-server" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.926419 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfa8b6b6-5e7b-4f5b-91b7-dc3dfb5ab4d3" containerName="registry-server" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.928205 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.934888 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.952918 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.953680 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs47z\" (UniqueName: \"kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:21 crc kubenswrapper[4867]: I1212 08:36:21.953760 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.055431 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.055612 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs47z\" (UniqueName: \"kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.055649 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.056571 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.056675 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.077148 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs47z\" (UniqueName: \"kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z\") pod \"community-operators-rbdbs\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.258315 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:22 crc kubenswrapper[4867]: I1212 08:36:22.803128 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:23 crc kubenswrapper[4867]: I1212 08:36:23.045381 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerStarted","Data":"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05"} Dec 12 08:36:23 crc kubenswrapper[4867]: I1212 08:36:23.045648 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerStarted","Data":"4ce77ddf6bafb0b610f59f24de945377df3aa77da0663c55e021ec0f25d7cd7c"} Dec 12 08:36:24 crc kubenswrapper[4867]: I1212 08:36:24.057491 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerID="dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05" exitCode=0 Dec 12 08:36:24 crc kubenswrapper[4867]: I1212 08:36:24.057556 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerDied","Data":"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05"} Dec 12 08:36:24 crc kubenswrapper[4867]: I1212 08:36:24.060036 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:36:25 crc kubenswrapper[4867]: I1212 08:36:25.072030 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerStarted","Data":"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a"} Dec 12 08:36:26 crc kubenswrapper[4867]: I1212 08:36:26.091760 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerID="3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a" exitCode=0 Dec 12 08:36:26 crc kubenswrapper[4867]: I1212 08:36:26.091804 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerDied","Data":"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a"} Dec 12 08:36:28 crc kubenswrapper[4867]: I1212 08:36:28.127904 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerStarted","Data":"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892"} Dec 12 08:36:28 crc kubenswrapper[4867]: I1212 08:36:28.163116 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rbdbs" podStartSLOduration=4.292994138 podStartE2EDuration="7.163078547s" podCreationTimestamp="2025-12-12 08:36:21 +0000 UTC" firstStartedPulling="2025-12-12 08:36:24.059829134 +0000 UTC m=+6471.631210403" lastFinishedPulling="2025-12-12 08:36:26.929913543 +0000 UTC m=+6474.501294812" observedRunningTime="2025-12-12 08:36:28.153886319 +0000 UTC m=+6475.725267598" watchObservedRunningTime="2025-12-12 08:36:28.163078547 +0000 UTC m=+6475.734459816" Dec 12 08:36:28 crc kubenswrapper[4867]: I1212 08:36:28.989340 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:36:28 crc kubenswrapper[4867]: I1212 08:36:28.989683 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:36:32 crc kubenswrapper[4867]: I1212 08:36:32.259153 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:32 crc kubenswrapper[4867]: I1212 08:36:32.259514 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:32 crc kubenswrapper[4867]: I1212 08:36:32.311759 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:33 crc kubenswrapper[4867]: I1212 08:36:33.217135 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:33 crc kubenswrapper[4867]: I1212 08:36:33.276941 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.183551 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rbdbs" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="registry-server" containerID="cri-o://4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892" gracePeriod=2 Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.636441 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.777268 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs47z\" (UniqueName: \"kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z\") pod \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.777386 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities\") pod \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.777487 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content\") pod \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\" (UID: \"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9\") " Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.778750 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities" (OuterVolumeSpecName: "utilities") pod "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" (UID: "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.782488 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z" (OuterVolumeSpecName: "kube-api-access-rs47z") pod "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" (UID: "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9"). InnerVolumeSpecName "kube-api-access-rs47z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.833193 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" (UID: "c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.879710 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.879740 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:36:35 crc kubenswrapper[4867]: I1212 08:36:35.879753 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs47z\" (UniqueName: \"kubernetes.io/projected/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9-kube-api-access-rs47z\") on node \"crc\" DevicePath \"\"" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.199552 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerID="4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892" exitCode=0 Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.199600 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerDied","Data":"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892"} Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.199621 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rbdbs" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.199643 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rbdbs" event={"ID":"c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9","Type":"ContainerDied","Data":"4ce77ddf6bafb0b610f59f24de945377df3aa77da0663c55e021ec0f25d7cd7c"} Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.199675 4867 scope.go:117] "RemoveContainer" containerID="4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.230637 4867 scope.go:117] "RemoveContainer" containerID="3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.238565 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.248964 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rbdbs"] Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.260532 4867 scope.go:117] "RemoveContainer" containerID="dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.315019 4867 scope.go:117] "RemoveContainer" containerID="4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892" Dec 12 08:36:36 crc kubenswrapper[4867]: E1212 08:36:36.315541 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892\": container with ID starting with 4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892 not found: ID does not exist" containerID="4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.315584 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892"} err="failed to get container status \"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892\": rpc error: code = NotFound desc = could not find container \"4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892\": container with ID starting with 4777a1975902108a12e6c37ebe9056cdab53db3bbe3da9f33ccce791d852e892 not found: ID does not exist" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.315610 4867 scope.go:117] "RemoveContainer" containerID="3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a" Dec 12 08:36:36 crc kubenswrapper[4867]: E1212 08:36:36.316557 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a\": container with ID starting with 3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a not found: ID does not exist" containerID="3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.316607 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a"} err="failed to get container status \"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a\": rpc error: code = NotFound desc = could not find container \"3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a\": container with ID starting with 3cd7022bed3229e91726dc36c8f83519f07365d4b57644e08df4dc2a683a709a not found: ID does not exist" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.316671 4867 scope.go:117] "RemoveContainer" containerID="dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05" Dec 12 08:36:36 crc kubenswrapper[4867]: E1212 08:36:36.317360 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05\": container with ID starting with dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05 not found: ID does not exist" containerID="dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.317391 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05"} err="failed to get container status \"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05\": rpc error: code = NotFound desc = could not find container \"dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05\": container with ID starting with dd785699cb910ecc77e090477a470dfc4f922b69bf582f26095d00048ce64f05 not found: ID does not exist" Dec 12 08:36:36 crc kubenswrapper[4867]: E1212 08:36:36.391454 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7d8a4a7_a97b_4900_a33b_eaf12c4c9ff9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7d8a4a7_a97b_4900_a33b_eaf12c4c9ff9.slice/crio-4ce77ddf6bafb0b610f59f24de945377df3aa77da0663c55e021ec0f25d7cd7c\": RecentStats: unable to find data in memory cache]" Dec 12 08:36:36 crc kubenswrapper[4867]: I1212 08:36:36.850447 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" path="/var/lib/kubelet/pods/c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9/volumes" Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.041594 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pnzm6"] Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.050060 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pnzm6"] Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.060222 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-2247-account-create-update-gtr66"] Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.073975 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-2247-account-create-update-gtr66"] Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.863195 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e" path="/var/lib/kubelet/pods/7d33ce8e-3c6a-40ef-b7e3-11e36b73eb3e/volumes" Dec 12 08:36:54 crc kubenswrapper[4867]: I1212 08:36:54.863997 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abaeb799-309d-4a35-9e9b-cdcfe0a4065f" path="/var/lib/kubelet/pods/abaeb799-309d-4a35-9e9b-cdcfe0a4065f/volumes" Dec 12 08:36:58 crc kubenswrapper[4867]: I1212 08:36:58.988946 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:36:58 crc kubenswrapper[4867]: I1212 08:36:58.989659 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:36:58 crc kubenswrapper[4867]: I1212 08:36:58.989706 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:36:58 crc kubenswrapper[4867]: I1212 08:36:58.990552 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:36:58 crc kubenswrapper[4867]: I1212 08:36:58.990613 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d" gracePeriod=600 Dec 12 08:36:59 crc kubenswrapper[4867]: I1212 08:36:59.429220 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d" exitCode=0 Dec 12 08:36:59 crc kubenswrapper[4867]: I1212 08:36:59.429338 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d"} Dec 12 08:36:59 crc kubenswrapper[4867]: I1212 08:36:59.429670 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35"} Dec 12 08:36:59 crc kubenswrapper[4867]: I1212 08:36:59.429695 4867 scope.go:117] "RemoveContainer" containerID="719e7a4d0fe652553c27186310e9df59cc25a448111cc0ae3664da8da782e098" Dec 12 08:37:01 crc kubenswrapper[4867]: I1212 08:37:01.817657 4867 scope.go:117] "RemoveContainer" containerID="e4e4d824a2401411622e7fb861727ad1415ec2046eaaf106644e9e3ec9a9f778" Dec 12 08:37:01 crc kubenswrapper[4867]: I1212 08:37:01.840345 4867 scope.go:117] "RemoveContainer" containerID="02430ea5aa82557393d2a4654b9ce856be20ad2afc41385b3e21992d93e523c5" Dec 12 08:37:08 crc kubenswrapper[4867]: I1212 08:37:08.043940 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-xgdqt"] Dec 12 08:37:08 crc kubenswrapper[4867]: I1212 08:37:08.055376 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-xgdqt"] Dec 12 08:37:08 crc kubenswrapper[4867]: I1212 08:37:08.854422 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fae1753-5101-4148-a0ca-20a48294a726" path="/var/lib/kubelet/pods/0fae1753-5101-4148-a0ca-20a48294a726/volumes" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.200880 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:37:50 crc kubenswrapper[4867]: E1212 08:37:50.201986 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="extract-utilities" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.202001 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="extract-utilities" Dec 12 08:37:50 crc kubenswrapper[4867]: E1212 08:37:50.202039 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="registry-server" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.202049 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="registry-server" Dec 12 08:37:50 crc kubenswrapper[4867]: E1212 08:37:50.202067 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="extract-content" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.202076 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="extract-content" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.202329 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d8a4a7-a97b-4900-a33b-eaf12c4c9ff9" containerName="registry-server" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.204061 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.209203 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.368475 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.368879 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdq4b\" (UniqueName: \"kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.369067 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.470565 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.470681 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdq4b\" (UniqueName: \"kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.470823 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.471111 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.471257 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.492121 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdq4b\" (UniqueName: \"kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b\") pod \"redhat-operators-gzd4q\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:50 crc kubenswrapper[4867]: I1212 08:37:50.535864 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:37:51 crc kubenswrapper[4867]: I1212 08:37:51.014946 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:37:51 crc kubenswrapper[4867]: I1212 08:37:51.946673 4867 generic.go:334] "Generic (PLEG): container finished" podID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerID="88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c" exitCode=0 Dec 12 08:37:51 crc kubenswrapper[4867]: I1212 08:37:51.946740 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerDied","Data":"88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c"} Dec 12 08:37:51 crc kubenswrapper[4867]: I1212 08:37:51.947024 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerStarted","Data":"d8e7f1f396fd04a489ad78367ae92ed52ef4bdca9bbf6c87c00bd68230d45147"} Dec 12 08:37:53 crc kubenswrapper[4867]: I1212 08:37:53.972941 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerStarted","Data":"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc"} Dec 12 08:37:56 crc kubenswrapper[4867]: I1212 08:37:56.999357 4867 generic.go:334] "Generic (PLEG): container finished" podID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerID="b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc" exitCode=0 Dec 12 08:37:56 crc kubenswrapper[4867]: I1212 08:37:56.999428 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerDied","Data":"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc"} Dec 12 08:37:58 crc kubenswrapper[4867]: I1212 08:37:58.010900 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerStarted","Data":"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8"} Dec 12 08:37:58 crc kubenswrapper[4867]: I1212 08:37:58.026271 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gzd4q" podStartSLOduration=2.440238522 podStartE2EDuration="8.026254254s" podCreationTimestamp="2025-12-12 08:37:50 +0000 UTC" firstStartedPulling="2025-12-12 08:37:51.948088582 +0000 UTC m=+6559.519469841" lastFinishedPulling="2025-12-12 08:37:57.534104304 +0000 UTC m=+6565.105485573" observedRunningTime="2025-12-12 08:37:58.025663418 +0000 UTC m=+6565.597044707" watchObservedRunningTime="2025-12-12 08:37:58.026254254 +0000 UTC m=+6565.597635523" Dec 12 08:38:00 crc kubenswrapper[4867]: I1212 08:38:00.536570 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:00 crc kubenswrapper[4867]: I1212 08:38:00.536932 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:01 crc kubenswrapper[4867]: I1212 08:38:01.582516 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gzd4q" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="registry-server" probeResult="failure" output=< Dec 12 08:38:01 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 08:38:01 crc kubenswrapper[4867]: > Dec 12 08:38:01 crc kubenswrapper[4867]: I1212 08:38:01.980093 4867 scope.go:117] "RemoveContainer" containerID="ef41feb6a82e9c2d1d78602fa8f55847064fe9a7d708b0d7fed7d52217855a01" Dec 12 08:38:10 crc kubenswrapper[4867]: I1212 08:38:10.586949 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:10 crc kubenswrapper[4867]: I1212 08:38:10.635066 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:10 crc kubenswrapper[4867]: I1212 08:38:10.820680 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.132380 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gzd4q" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="registry-server" containerID="cri-o://b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8" gracePeriod=2 Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.641903 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.743618 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities\") pod \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.743753 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdq4b\" (UniqueName: \"kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b\") pod \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.743803 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content\") pod \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\" (UID: \"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4\") " Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.744964 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities" (OuterVolumeSpecName: "utilities") pod "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" (UID: "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.756004 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b" (OuterVolumeSpecName: "kube-api-access-gdq4b") pod "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" (UID: "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4"). InnerVolumeSpecName "kube-api-access-gdq4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.845788 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.845820 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdq4b\" (UniqueName: \"kubernetes.io/projected/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-kube-api-access-gdq4b\") on node \"crc\" DevicePath \"\"" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.881801 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" (UID: "37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:38:12 crc kubenswrapper[4867]: I1212 08:38:12.950762 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.142903 4867 generic.go:334] "Generic (PLEG): container finished" podID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerID="b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8" exitCode=0 Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.142946 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerDied","Data":"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8"} Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.142972 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gzd4q" event={"ID":"37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4","Type":"ContainerDied","Data":"d8e7f1f396fd04a489ad78367ae92ed52ef4bdca9bbf6c87c00bd68230d45147"} Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.142989 4867 scope.go:117] "RemoveContainer" containerID="b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.143004 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gzd4q" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.177206 4867 scope.go:117] "RemoveContainer" containerID="b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.181036 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.191448 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gzd4q"] Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.206718 4867 scope.go:117] "RemoveContainer" containerID="88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.262420 4867 scope.go:117] "RemoveContainer" containerID="b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8" Dec 12 08:38:13 crc kubenswrapper[4867]: E1212 08:38:13.263047 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8\": container with ID starting with b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8 not found: ID does not exist" containerID="b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.263107 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8"} err="failed to get container status \"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8\": rpc error: code = NotFound desc = could not find container \"b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8\": container with ID starting with b898ed8951f7dfbc193a14438dccf85db19f2c610a5b1fe92d6f71258d0950b8 not found: ID does not exist" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.263145 4867 scope.go:117] "RemoveContainer" containerID="b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc" Dec 12 08:38:13 crc kubenswrapper[4867]: E1212 08:38:13.264252 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc\": container with ID starting with b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc not found: ID does not exist" containerID="b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.264308 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc"} err="failed to get container status \"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc\": rpc error: code = NotFound desc = could not find container \"b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc\": container with ID starting with b16ccca3a718a01682b0a5c16204e03bd9555bc354245c3a2f6c5e7b181c4bbc not found: ID does not exist" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.264339 4867 scope.go:117] "RemoveContainer" containerID="88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c" Dec 12 08:38:13 crc kubenswrapper[4867]: E1212 08:38:13.264689 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c\": container with ID starting with 88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c not found: ID does not exist" containerID="88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c" Dec 12 08:38:13 crc kubenswrapper[4867]: I1212 08:38:13.264723 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c"} err="failed to get container status \"88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c\": rpc error: code = NotFound desc = could not find container \"88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c\": container with ID starting with 88f586b0c382df52e147997d36d8ca982581111806be1463b34f83e43ee1d71c not found: ID does not exist" Dec 12 08:38:14 crc kubenswrapper[4867]: I1212 08:38:14.854711 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" path="/var/lib/kubelet/pods/37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4/volumes" Dec 12 08:39:28 crc kubenswrapper[4867]: I1212 08:39:28.989391 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:39:28 crc kubenswrapper[4867]: I1212 08:39:28.990052 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:39:41 crc kubenswrapper[4867]: I1212 08:39:41.046183 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-m6nfp"] Dec 12 08:39:41 crc kubenswrapper[4867]: I1212 08:39:41.057870 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-5322-account-create-update-bql4k"] Dec 12 08:39:41 crc kubenswrapper[4867]: I1212 08:39:41.067807 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-m6nfp"] Dec 12 08:39:41 crc kubenswrapper[4867]: I1212 08:39:41.077013 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-5322-account-create-update-bql4k"] Dec 12 08:39:42 crc kubenswrapper[4867]: I1212 08:39:42.852901 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93" path="/var/lib/kubelet/pods/0535ef1e-ae2b-44b6-92d3-fc5e7a3aeb93/volumes" Dec 12 08:39:42 crc kubenswrapper[4867]: I1212 08:39:42.854304 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c258b557-6d4e-4ba6-b356-1a716ca6cd3b" path="/var/lib/kubelet/pods/c258b557-6d4e-4ba6-b356-1a716ca6cd3b/volumes" Dec 12 08:39:54 crc kubenswrapper[4867]: I1212 08:39:54.043722 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-8bq7b"] Dec 12 08:39:54 crc kubenswrapper[4867]: I1212 08:39:54.058977 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-8bq7b"] Dec 12 08:39:54 crc kubenswrapper[4867]: I1212 08:39:54.849683 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39a39479-d1b6-45a2-a480-c15c61be7627" path="/var/lib/kubelet/pods/39a39479-d1b6-45a2-a480-c15c61be7627/volumes" Dec 12 08:39:58 crc kubenswrapper[4867]: I1212 08:39:58.989300 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:39:58 crc kubenswrapper[4867]: I1212 08:39:58.989955 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:40:02 crc kubenswrapper[4867]: I1212 08:40:02.076797 4867 scope.go:117] "RemoveContainer" containerID="e952512f990fb58145146f8d3812dc427af9c69fe0b57c73ca6e72026b02d963" Dec 12 08:40:02 crc kubenswrapper[4867]: I1212 08:40:02.107715 4867 scope.go:117] "RemoveContainer" containerID="b4f4de8aa84fe727eade28ae8bfa7516b2a12a2038800ef6d5415ea12c79db54" Dec 12 08:40:02 crc kubenswrapper[4867]: I1212 08:40:02.157744 4867 scope.go:117] "RemoveContainer" containerID="46d38b6eb0e540ba5fbd73f8bf7407d6c2ddfdcee85ad1e0d27374e681efa3a5" Dec 12 08:40:28 crc kubenswrapper[4867]: I1212 08:40:28.989155 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:40:28 crc kubenswrapper[4867]: I1212 08:40:28.989753 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:40:28 crc kubenswrapper[4867]: I1212 08:40:28.989796 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:40:28 crc kubenswrapper[4867]: I1212 08:40:28.990632 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:40:28 crc kubenswrapper[4867]: I1212 08:40:28.990691 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" gracePeriod=600 Dec 12 08:40:29 crc kubenswrapper[4867]: E1212 08:40:29.108693 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:40:29 crc kubenswrapper[4867]: I1212 08:40:29.387587 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" exitCode=0 Dec 12 08:40:29 crc kubenswrapper[4867]: I1212 08:40:29.387640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35"} Dec 12 08:40:29 crc kubenswrapper[4867]: I1212 08:40:29.387686 4867 scope.go:117] "RemoveContainer" containerID="b857014c45b0447bad67753e6e730f810ad75d8d1e0824453f99791b17dccf9d" Dec 12 08:40:29 crc kubenswrapper[4867]: I1212 08:40:29.388845 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:40:29 crc kubenswrapper[4867]: E1212 08:40:29.389277 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:40:43 crc kubenswrapper[4867]: I1212 08:40:43.839551 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:40:43 crc kubenswrapper[4867]: E1212 08:40:43.840442 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:40:58 crc kubenswrapper[4867]: I1212 08:40:58.837845 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:40:58 crc kubenswrapper[4867]: E1212 08:40:58.838680 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:41:13 crc kubenswrapper[4867]: I1212 08:41:13.838364 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:41:13 crc kubenswrapper[4867]: E1212 08:41:13.839009 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:41:26 crc kubenswrapper[4867]: I1212 08:41:26.838262 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:41:26 crc kubenswrapper[4867]: E1212 08:41:26.839062 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:41:39 crc kubenswrapper[4867]: I1212 08:41:39.837921 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:41:39 crc kubenswrapper[4867]: E1212 08:41:39.838710 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:41:53 crc kubenswrapper[4867]: I1212 08:41:53.838525 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:41:53 crc kubenswrapper[4867]: E1212 08:41:53.839638 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:41:54 crc kubenswrapper[4867]: E1212 08:41:54.910189 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55c56e88_ce42_4abb_bdfb_e68fd154fea6.slice/crio-24d2676e5cc5576ee28d4d9fb8b71a1e290498b3cd6d2a9cdd701926d35228c8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55c56e88_ce42_4abb_bdfb_e68fd154fea6.slice/crio-conmon-24d2676e5cc5576ee28d4d9fb8b71a1e290498b3cd6d2a9cdd701926d35228c8.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:41:55 crc kubenswrapper[4867]: I1212 08:41:55.192966 4867 generic.go:334] "Generic (PLEG): container finished" podID="55c56e88-ce42-4abb-bdfb-e68fd154fea6" containerID="24d2676e5cc5576ee28d4d9fb8b71a1e290498b3cd6d2a9cdd701926d35228c8" exitCode=0 Dec 12 08:41:55 crc kubenswrapper[4867]: I1212 08:41:55.193184 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" event={"ID":"55c56e88-ce42-4abb-bdfb-e68fd154fea6","Type":"ContainerDied","Data":"24d2676e5cc5576ee28d4d9fb8b71a1e290498b3cd6d2a9cdd701926d35228c8"} Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.665712 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.805653 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory\") pod \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.805828 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle\") pod \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.805921 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc\") pod \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.806020 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key\") pod \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\" (UID: \"55c56e88-ce42-4abb-bdfb-e68fd154fea6\") " Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.811240 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "55c56e88-ce42-4abb-bdfb-e68fd154fea6" (UID: "55c56e88-ce42-4abb-bdfb-e68fd154fea6"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.811286 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc" (OuterVolumeSpecName: "kube-api-access-47lhc") pod "55c56e88-ce42-4abb-bdfb-e68fd154fea6" (UID: "55c56e88-ce42-4abb-bdfb-e68fd154fea6"). InnerVolumeSpecName "kube-api-access-47lhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.833908 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory" (OuterVolumeSpecName: "inventory") pod "55c56e88-ce42-4abb-bdfb-e68fd154fea6" (UID: "55c56e88-ce42-4abb-bdfb-e68fd154fea6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.838880 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "55c56e88-ce42-4abb-bdfb-e68fd154fea6" (UID: "55c56e88-ce42-4abb-bdfb-e68fd154fea6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.908827 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.908858 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.908870 4867 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55c56e88-ce42-4abb-bdfb-e68fd154fea6-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:41:56 crc kubenswrapper[4867]: I1212 08:41:56.908885 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47lhc\" (UniqueName: \"kubernetes.io/projected/55c56e88-ce42-4abb-bdfb-e68fd154fea6-kube-api-access-47lhc\") on node \"crc\" DevicePath \"\"" Dec 12 08:41:57 crc kubenswrapper[4867]: I1212 08:41:57.213041 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" event={"ID":"55c56e88-ce42-4abb-bdfb-e68fd154fea6","Type":"ContainerDied","Data":"dcbed3c754a7675577e6ed09ef6f17ea1911e4905f19f5a21f5c563f1efbcc3b"} Dec 12 08:41:57 crc kubenswrapper[4867]: I1212 08:41:57.213092 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcbed3c754a7675577e6ed09ef6f17ea1911e4905f19f5a21f5c563f1efbcc3b" Dec 12 08:41:57 crc kubenswrapper[4867]: I1212 08:41:57.213183 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.345659 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-q2znc"] Dec 12 08:42:06 crc kubenswrapper[4867]: E1212 08:42:06.346470 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="registry-server" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.346483 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="registry-server" Dec 12 08:42:06 crc kubenswrapper[4867]: E1212 08:42:06.346501 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="extract-utilities" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.346507 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="extract-utilities" Dec 12 08:42:06 crc kubenswrapper[4867]: E1212 08:42:06.346520 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55c56e88-ce42-4abb-bdfb-e68fd154fea6" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.346527 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="55c56e88-ce42-4abb-bdfb-e68fd154fea6" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 12 08:42:06 crc kubenswrapper[4867]: E1212 08:42:06.346542 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="extract-content" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.346548 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="extract-content" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.347555 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e9e4f5-cfd5-4cc6-a07c-b389eef40fa4" containerName="registry-server" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.347573 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="55c56e88-ce42-4abb-bdfb-e68fd154fea6" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.348237 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.350983 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.356147 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.356428 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.356605 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.395412 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-q2znc"] Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.506281 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.506524 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.506647 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.506718 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzqv2\" (UniqueName: \"kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.608674 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.608755 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.608810 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.608838 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzqv2\" (UniqueName: \"kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.615111 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.617732 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.618346 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.628108 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzqv2\" (UniqueName: \"kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2\") pod \"bootstrap-openstack-openstack-cell1-q2znc\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.670495 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:42:06 crc kubenswrapper[4867]: I1212 08:42:06.837998 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:42:06 crc kubenswrapper[4867]: E1212 08:42:06.838526 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:42:07 crc kubenswrapper[4867]: I1212 08:42:07.186523 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-q2znc"] Dec 12 08:42:07 crc kubenswrapper[4867]: I1212 08:42:07.190491 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:42:07 crc kubenswrapper[4867]: I1212 08:42:07.318701 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" event={"ID":"53b9ceda-53ce-4426-9075-83f190cc8351","Type":"ContainerStarted","Data":"9531149b99e15c6ee8d8b24c25d3691031b21c5da97c2ca7e5924cc9a0bad371"} Dec 12 08:42:08 crc kubenswrapper[4867]: I1212 08:42:08.329610 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" event={"ID":"53b9ceda-53ce-4426-9075-83f190cc8351","Type":"ContainerStarted","Data":"0b9f6685d3f2ccaa46573030a13cc80d7d7995927a9be8ded205a986d9af4c07"} Dec 12 08:42:08 crc kubenswrapper[4867]: I1212 08:42:08.351372 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" podStartSLOduration=1.745291521 podStartE2EDuration="2.351353334s" podCreationTimestamp="2025-12-12 08:42:06 +0000 UTC" firstStartedPulling="2025-12-12 08:42:07.190191694 +0000 UTC m=+6814.761572973" lastFinishedPulling="2025-12-12 08:42:07.796253517 +0000 UTC m=+6815.367634786" observedRunningTime="2025-12-12 08:42:08.34290674 +0000 UTC m=+6815.914288019" watchObservedRunningTime="2025-12-12 08:42:08.351353334 +0000 UTC m=+6815.922734603" Dec 12 08:42:21 crc kubenswrapper[4867]: I1212 08:42:21.838998 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:42:21 crc kubenswrapper[4867]: E1212 08:42:21.840166 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.403784 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.406682 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.414947 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.593087 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.593691 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tnjc\" (UniqueName: \"kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.593823 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.695873 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.696001 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.696059 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tnjc\" (UniqueName: \"kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.696463 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.696557 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.722764 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tnjc\" (UniqueName: \"kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc\") pod \"redhat-marketplace-vkmjb\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:33 crc kubenswrapper[4867]: I1212 08:42:33.728690 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:34 crc kubenswrapper[4867]: I1212 08:42:34.207720 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:34 crc kubenswrapper[4867]: I1212 08:42:34.562643 4867 generic.go:334] "Generic (PLEG): container finished" podID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerID="ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614" exitCode=0 Dec 12 08:42:34 crc kubenswrapper[4867]: I1212 08:42:34.562714 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerDied","Data":"ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614"} Dec 12 08:42:34 crc kubenswrapper[4867]: I1212 08:42:34.563011 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerStarted","Data":"71ad7f58f871f3073dbb386c8dcf0fa094e629b318f8469c2be811c694b0eea7"} Dec 12 08:42:36 crc kubenswrapper[4867]: I1212 08:42:36.578334 4867 generic.go:334] "Generic (PLEG): container finished" podID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerID="c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac" exitCode=0 Dec 12 08:42:36 crc kubenswrapper[4867]: I1212 08:42:36.578836 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerDied","Data":"c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac"} Dec 12 08:42:36 crc kubenswrapper[4867]: I1212 08:42:36.838486 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:42:36 crc kubenswrapper[4867]: E1212 08:42:36.838740 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:42:37 crc kubenswrapper[4867]: I1212 08:42:37.603774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerStarted","Data":"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72"} Dec 12 08:42:37 crc kubenswrapper[4867]: I1212 08:42:37.638851 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vkmjb" podStartSLOduration=2.025401748 podStartE2EDuration="4.638821155s" podCreationTimestamp="2025-12-12 08:42:33 +0000 UTC" firstStartedPulling="2025-12-12 08:42:34.566282514 +0000 UTC m=+6842.137663783" lastFinishedPulling="2025-12-12 08:42:37.179701881 +0000 UTC m=+6844.751083190" observedRunningTime="2025-12-12 08:42:37.625282287 +0000 UTC m=+6845.196663646" watchObservedRunningTime="2025-12-12 08:42:37.638821155 +0000 UTC m=+6845.210202464" Dec 12 08:42:43 crc kubenswrapper[4867]: I1212 08:42:43.729424 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:43 crc kubenswrapper[4867]: I1212 08:42:43.730060 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:43 crc kubenswrapper[4867]: I1212 08:42:43.772780 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:44 crc kubenswrapper[4867]: I1212 08:42:44.709123 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:44 crc kubenswrapper[4867]: I1212 08:42:44.755164 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:46 crc kubenswrapper[4867]: I1212 08:42:46.686550 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vkmjb" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="registry-server" containerID="cri-o://9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72" gracePeriod=2 Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.136544 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.268980 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content\") pod \"e1f1991b-eaa7-4c77-9b00-622da77fb255\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.269282 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities\") pod \"e1f1991b-eaa7-4c77-9b00-622da77fb255\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.269395 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tnjc\" (UniqueName: \"kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc\") pod \"e1f1991b-eaa7-4c77-9b00-622da77fb255\" (UID: \"e1f1991b-eaa7-4c77-9b00-622da77fb255\") " Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.270595 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities" (OuterVolumeSpecName: "utilities") pod "e1f1991b-eaa7-4c77-9b00-622da77fb255" (UID: "e1f1991b-eaa7-4c77-9b00-622da77fb255"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.270696 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.278889 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc" (OuterVolumeSpecName: "kube-api-access-9tnjc") pod "e1f1991b-eaa7-4c77-9b00-622da77fb255" (UID: "e1f1991b-eaa7-4c77-9b00-622da77fb255"). InnerVolumeSpecName "kube-api-access-9tnjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.294377 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1f1991b-eaa7-4c77-9b00-622da77fb255" (UID: "e1f1991b-eaa7-4c77-9b00-622da77fb255"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.373187 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f1991b-eaa7-4c77-9b00-622da77fb255-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.373250 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tnjc\" (UniqueName: \"kubernetes.io/projected/e1f1991b-eaa7-4c77-9b00-622da77fb255-kube-api-access-9tnjc\") on node \"crc\" DevicePath \"\"" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.700336 4867 generic.go:334] "Generic (PLEG): container finished" podID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerID="9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72" exitCode=0 Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.700382 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkmjb" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.700401 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerDied","Data":"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72"} Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.700440 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkmjb" event={"ID":"e1f1991b-eaa7-4c77-9b00-622da77fb255","Type":"ContainerDied","Data":"71ad7f58f871f3073dbb386c8dcf0fa094e629b318f8469c2be811c694b0eea7"} Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.700469 4867 scope.go:117] "RemoveContainer" containerID="9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.735356 4867 scope.go:117] "RemoveContainer" containerID="c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.739285 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.747303 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkmjb"] Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.782720 4867 scope.go:117] "RemoveContainer" containerID="ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.824136 4867 scope.go:117] "RemoveContainer" containerID="9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72" Dec 12 08:42:47 crc kubenswrapper[4867]: E1212 08:42:47.825126 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72\": container with ID starting with 9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72 not found: ID does not exist" containerID="9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.825200 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72"} err="failed to get container status \"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72\": rpc error: code = NotFound desc = could not find container \"9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72\": container with ID starting with 9b4e537823123f97d785906e49757afdea148882a93532cff5ae4c406a9a3a72 not found: ID does not exist" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.825248 4867 scope.go:117] "RemoveContainer" containerID="c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac" Dec 12 08:42:47 crc kubenswrapper[4867]: E1212 08:42:47.825799 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac\": container with ID starting with c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac not found: ID does not exist" containerID="c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.825872 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac"} err="failed to get container status \"c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac\": rpc error: code = NotFound desc = could not find container \"c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac\": container with ID starting with c497c97ac9814b79e50f85ddd6c8761ba1cffa8255d8004631a5b18ee9d900ac not found: ID does not exist" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.825922 4867 scope.go:117] "RemoveContainer" containerID="ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614" Dec 12 08:42:47 crc kubenswrapper[4867]: E1212 08:42:47.826395 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614\": container with ID starting with ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614 not found: ID does not exist" containerID="ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614" Dec 12 08:42:47 crc kubenswrapper[4867]: I1212 08:42:47.826416 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614"} err="failed to get container status \"ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614\": rpc error: code = NotFound desc = could not find container \"ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614\": container with ID starting with ee38cff01f8787e68fe21c2cbbc6c41945838bbdc3f0a63d7bec8e7db0b37614 not found: ID does not exist" Dec 12 08:42:48 crc kubenswrapper[4867]: I1212 08:42:48.850642 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" path="/var/lib/kubelet/pods/e1f1991b-eaa7-4c77-9b00-622da77fb255/volumes" Dec 12 08:42:51 crc kubenswrapper[4867]: I1212 08:42:51.838483 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:42:51 crc kubenswrapper[4867]: E1212 08:42:51.839010 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:43:06 crc kubenswrapper[4867]: I1212 08:43:06.838563 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:43:06 crc kubenswrapper[4867]: E1212 08:43:06.839188 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:43:20 crc kubenswrapper[4867]: I1212 08:43:20.838705 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:43:20 crc kubenswrapper[4867]: E1212 08:43:20.839685 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:43:33 crc kubenswrapper[4867]: I1212 08:43:33.838943 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:43:33 crc kubenswrapper[4867]: E1212 08:43:33.839807 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:43:44 crc kubenswrapper[4867]: I1212 08:43:44.838109 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:43:44 crc kubenswrapper[4867]: E1212 08:43:44.838933 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:43:59 crc kubenswrapper[4867]: I1212 08:43:59.838278 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:43:59 crc kubenswrapper[4867]: E1212 08:43:59.839157 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:44:11 crc kubenswrapper[4867]: I1212 08:44:11.837945 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:44:11 crc kubenswrapper[4867]: E1212 08:44:11.838719 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:44:22 crc kubenswrapper[4867]: I1212 08:44:22.851356 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:44:22 crc kubenswrapper[4867]: E1212 08:44:22.852268 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:44:34 crc kubenswrapper[4867]: I1212 08:44:34.838208 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:44:34 crc kubenswrapper[4867]: E1212 08:44:34.839051 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:44:46 crc kubenswrapper[4867]: I1212 08:44:46.839562 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:44:46 crc kubenswrapper[4867]: E1212 08:44:46.840488 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.242606 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch"] Dec 12 08:45:00 crc kubenswrapper[4867]: E1212 08:45:00.244317 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="registry-server" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.244339 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="registry-server" Dec 12 08:45:00 crc kubenswrapper[4867]: E1212 08:45:00.244375 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="extract-utilities" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.244383 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="extract-utilities" Dec 12 08:45:00 crc kubenswrapper[4867]: E1212 08:45:00.244416 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="extract-content" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.244559 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="extract-content" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.246586 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f1991b-eaa7-4c77-9b00-622da77fb255" containerName="registry-server" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.262025 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.284153 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.301508 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.308664 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch"] Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.366672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.366806 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.366852 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vd4v\" (UniqueName: \"kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.468620 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.469120 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.469167 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vd4v\" (UniqueName: \"kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.469687 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.479314 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.484784 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vd4v\" (UniqueName: \"kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v\") pod \"collect-profiles-29425485-mv6ch\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.674855 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:00 crc kubenswrapper[4867]: I1212 08:45:00.838871 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:45:00 crc kubenswrapper[4867]: E1212 08:45:00.839401 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:45:01 crc kubenswrapper[4867]: I1212 08:45:01.136795 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch"] Dec 12 08:45:02 crc kubenswrapper[4867]: I1212 08:45:02.017444 4867 generic.go:334] "Generic (PLEG): container finished" podID="fd03ed01-4350-4edc-8144-0e9d4175e9e5" containerID="050a4e7bc0da784d8490bb4698388fe14c6de9134c272058fc4dbae19ba488c9" exitCode=0 Dec 12 08:45:02 crc kubenswrapper[4867]: I1212 08:45:02.017507 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" event={"ID":"fd03ed01-4350-4edc-8144-0e9d4175e9e5","Type":"ContainerDied","Data":"050a4e7bc0da784d8490bb4698388fe14c6de9134c272058fc4dbae19ba488c9"} Dec 12 08:45:02 crc kubenswrapper[4867]: I1212 08:45:02.017752 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" event={"ID":"fd03ed01-4350-4edc-8144-0e9d4175e9e5","Type":"ContainerStarted","Data":"903437b1a5478a9bfaa6296c4311ce79c581cbb173f3d529a2ce759cd5c95876"} Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.439538 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.464088 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume\") pod \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.464160 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume\") pod \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.464206 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vd4v\" (UniqueName: \"kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v\") pod \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\" (UID: \"fd03ed01-4350-4edc-8144-0e9d4175e9e5\") " Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.465190 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "fd03ed01-4350-4edc-8144-0e9d4175e9e5" (UID: "fd03ed01-4350-4edc-8144-0e9d4175e9e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.470409 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fd03ed01-4350-4edc-8144-0e9d4175e9e5" (UID: "fd03ed01-4350-4edc-8144-0e9d4175e9e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.475401 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v" (OuterVolumeSpecName: "kube-api-access-4vd4v") pod "fd03ed01-4350-4edc-8144-0e9d4175e9e5" (UID: "fd03ed01-4350-4edc-8144-0e9d4175e9e5"). InnerVolumeSpecName "kube-api-access-4vd4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.565884 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd03ed01-4350-4edc-8144-0e9d4175e9e5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.565921 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd03ed01-4350-4edc-8144-0e9d4175e9e5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:03 crc kubenswrapper[4867]: I1212 08:45:03.565931 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vd4v\" (UniqueName: \"kubernetes.io/projected/fd03ed01-4350-4edc-8144-0e9d4175e9e5-kube-api-access-4vd4v\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.038066 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" event={"ID":"fd03ed01-4350-4edc-8144-0e9d4175e9e5","Type":"ContainerDied","Data":"903437b1a5478a9bfaa6296c4311ce79c581cbb173f3d529a2ce759cd5c95876"} Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.038409 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="903437b1a5478a9bfaa6296c4311ce79c581cbb173f3d529a2ce759cd5c95876" Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.038113 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch" Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.508204 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks"] Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.517408 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425440-694ks"] Dec 12 08:45:04 crc kubenswrapper[4867]: I1212 08:45:04.848907 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3487e86b-ff16-453e-8eb5-ce86d6491ab1" path="/var/lib/kubelet/pods/3487e86b-ff16-453e-8eb5-ce86d6491ab1/volumes" Dec 12 08:45:12 crc kubenswrapper[4867]: I1212 08:45:12.142777 4867 generic.go:334] "Generic (PLEG): container finished" podID="53b9ceda-53ce-4426-9075-83f190cc8351" containerID="0b9f6685d3f2ccaa46573030a13cc80d7d7995927a9be8ded205a986d9af4c07" exitCode=0 Dec 12 08:45:12 crc kubenswrapper[4867]: I1212 08:45:12.142837 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" event={"ID":"53b9ceda-53ce-4426-9075-83f190cc8351","Type":"ContainerDied","Data":"0b9f6685d3f2ccaa46573030a13cc80d7d7995927a9be8ded205a986d9af4c07"} Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.670908 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.855721 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key\") pod \"53b9ceda-53ce-4426-9075-83f190cc8351\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.855945 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle\") pod \"53b9ceda-53ce-4426-9075-83f190cc8351\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.855982 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzqv2\" (UniqueName: \"kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2\") pod \"53b9ceda-53ce-4426-9075-83f190cc8351\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.856092 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory\") pod \"53b9ceda-53ce-4426-9075-83f190cc8351\" (UID: \"53b9ceda-53ce-4426-9075-83f190cc8351\") " Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.862000 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2" (OuterVolumeSpecName: "kube-api-access-lzqv2") pod "53b9ceda-53ce-4426-9075-83f190cc8351" (UID: "53b9ceda-53ce-4426-9075-83f190cc8351"). InnerVolumeSpecName "kube-api-access-lzqv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.863853 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "53b9ceda-53ce-4426-9075-83f190cc8351" (UID: "53b9ceda-53ce-4426-9075-83f190cc8351"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.886127 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53b9ceda-53ce-4426-9075-83f190cc8351" (UID: "53b9ceda-53ce-4426-9075-83f190cc8351"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.888644 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory" (OuterVolumeSpecName: "inventory") pod "53b9ceda-53ce-4426-9075-83f190cc8351" (UID: "53b9ceda-53ce-4426-9075-83f190cc8351"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.958207 4867 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.958529 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzqv2\" (UniqueName: \"kubernetes.io/projected/53b9ceda-53ce-4426-9075-83f190cc8351-kube-api-access-lzqv2\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.958604 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:13 crc kubenswrapper[4867]: I1212 08:45:13.958677 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53b9ceda-53ce-4426-9075-83f190cc8351-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.161331 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" event={"ID":"53b9ceda-53ce-4426-9075-83f190cc8351","Type":"ContainerDied","Data":"9531149b99e15c6ee8d8b24c25d3691031b21c5da97c2ca7e5924cc9a0bad371"} Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.161393 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9531149b99e15c6ee8d8b24c25d3691031b21c5da97c2ca7e5924cc9a0bad371" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.161407 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-q2znc" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.253386 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-lfg7r"] Dec 12 08:45:14 crc kubenswrapper[4867]: E1212 08:45:14.254516 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd03ed01-4350-4edc-8144-0e9d4175e9e5" containerName="collect-profiles" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.254531 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd03ed01-4350-4edc-8144-0e9d4175e9e5" containerName="collect-profiles" Dec 12 08:45:14 crc kubenswrapper[4867]: E1212 08:45:14.254574 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b9ceda-53ce-4426-9075-83f190cc8351" containerName="bootstrap-openstack-openstack-cell1" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.254581 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b9ceda-53ce-4426-9075-83f190cc8351" containerName="bootstrap-openstack-openstack-cell1" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.254787 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd03ed01-4350-4edc-8144-0e9d4175e9e5" containerName="collect-profiles" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.254811 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b9ceda-53ce-4426-9075-83f190cc8351" containerName="bootstrap-openstack-openstack-cell1" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.255754 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.258021 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.258180 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.261024 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.261536 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.263825 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxk6w\" (UniqueName: \"kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.263894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.263931 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.264282 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-lfg7r"] Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.366211 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.366285 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.366453 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxk6w\" (UniqueName: \"kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.369794 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.369850 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.381447 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxk6w\" (UniqueName: \"kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w\") pod \"download-cache-openstack-openstack-cell1-lfg7r\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:14 crc kubenswrapper[4867]: I1212 08:45:14.572281 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:45:15 crc kubenswrapper[4867]: I1212 08:45:15.100416 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-lfg7r"] Dec 12 08:45:15 crc kubenswrapper[4867]: W1212 08:45:15.103658 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod146d44cb_ac9b_481c_8aa5_0884c7025d47.slice/crio-604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137 WatchSource:0}: Error finding container 604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137: Status 404 returned error can't find the container with id 604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137 Dec 12 08:45:15 crc kubenswrapper[4867]: I1212 08:45:15.172541 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" event={"ID":"146d44cb-ac9b-481c-8aa5-0884c7025d47","Type":"ContainerStarted","Data":"604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137"} Dec 12 08:45:15 crc kubenswrapper[4867]: I1212 08:45:15.838538 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:45:15 crc kubenswrapper[4867]: E1212 08:45:15.839069 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:45:16 crc kubenswrapper[4867]: I1212 08:45:16.182217 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" event={"ID":"146d44cb-ac9b-481c-8aa5-0884c7025d47","Type":"ContainerStarted","Data":"f7b123d83d0595f1d274ec35d7b007a111c8ade723144321d6f486ca7847416f"} Dec 12 08:45:16 crc kubenswrapper[4867]: I1212 08:45:16.208544 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" podStartSLOduration=1.7173842929999998 podStartE2EDuration="2.208527969s" podCreationTimestamp="2025-12-12 08:45:14 +0000 UTC" firstStartedPulling="2025-12-12 08:45:15.105748154 +0000 UTC m=+7002.677129423" lastFinishedPulling="2025-12-12 08:45:15.59689183 +0000 UTC m=+7003.168273099" observedRunningTime="2025-12-12 08:45:16.203942936 +0000 UTC m=+7003.775324205" watchObservedRunningTime="2025-12-12 08:45:16.208527969 +0000 UTC m=+7003.779909238" Dec 12 08:45:27 crc kubenswrapper[4867]: I1212 08:45:27.838332 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:45:27 crc kubenswrapper[4867]: E1212 08:45:27.839292 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:45:40 crc kubenswrapper[4867]: I1212 08:45:40.840083 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:45:41 crc kubenswrapper[4867]: I1212 08:45:41.421354 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2"} Dec 12 08:46:02 crc kubenswrapper[4867]: I1212 08:46:02.380933 4867 scope.go:117] "RemoveContainer" containerID="fff20affae3ac5e445bea8644333ece80900a81621c10fbb5152298c9b1b51cc" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.397262 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.403566 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.405254 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.484528 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.484762 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.484832 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5dq\" (UniqueName: \"kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.586583 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.587033 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.587077 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5dq\" (UniqueName: \"kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.587144 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.587504 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.616921 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5dq\" (UniqueName: \"kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq\") pod \"certified-operators-kddr7\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:03 crc kubenswrapper[4867]: I1212 08:46:03.725584 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:04 crc kubenswrapper[4867]: I1212 08:46:04.056877 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:04 crc kubenswrapper[4867]: I1212 08:46:04.652359 4867 generic.go:334] "Generic (PLEG): container finished" podID="851c35e9-4d8f-4062-bbf2-24340effc837" containerID="abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4" exitCode=0 Dec 12 08:46:04 crc kubenswrapper[4867]: I1212 08:46:04.652440 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerDied","Data":"abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4"} Dec 12 08:46:04 crc kubenswrapper[4867]: I1212 08:46:04.652747 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerStarted","Data":"b0ce08617bc973d617dbe97a980d1ea1aaa6da2433a4cff9be68606ad2872991"} Dec 12 08:46:05 crc kubenswrapper[4867]: I1212 08:46:05.663527 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerStarted","Data":"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3"} Dec 12 08:46:06 crc kubenswrapper[4867]: I1212 08:46:06.673212 4867 generic.go:334] "Generic (PLEG): container finished" podID="851c35e9-4d8f-4062-bbf2-24340effc837" containerID="dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3" exitCode=0 Dec 12 08:46:06 crc kubenswrapper[4867]: I1212 08:46:06.673293 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerDied","Data":"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3"} Dec 12 08:46:07 crc kubenswrapper[4867]: I1212 08:46:07.684823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerStarted","Data":"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006"} Dec 12 08:46:07 crc kubenswrapper[4867]: I1212 08:46:07.716765 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kddr7" podStartSLOduration=2.214629935 podStartE2EDuration="4.716743026s" podCreationTimestamp="2025-12-12 08:46:03 +0000 UTC" firstStartedPulling="2025-12-12 08:46:04.654809365 +0000 UTC m=+7052.226190634" lastFinishedPulling="2025-12-12 08:46:07.156922456 +0000 UTC m=+7054.728303725" observedRunningTime="2025-12-12 08:46:07.704128454 +0000 UTC m=+7055.275509743" watchObservedRunningTime="2025-12-12 08:46:07.716743026 +0000 UTC m=+7055.288124295" Dec 12 08:46:13 crc kubenswrapper[4867]: I1212 08:46:13.726329 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:13 crc kubenswrapper[4867]: I1212 08:46:13.726931 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:13 crc kubenswrapper[4867]: I1212 08:46:13.790340 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:14 crc kubenswrapper[4867]: I1212 08:46:14.794062 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:14 crc kubenswrapper[4867]: I1212 08:46:14.857117 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:16 crc kubenswrapper[4867]: I1212 08:46:16.763055 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kddr7" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="registry-server" containerID="cri-o://1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006" gracePeriod=2 Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.249259 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.453360 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content\") pod \"851c35e9-4d8f-4062-bbf2-24340effc837\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.453458 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities\") pod \"851c35e9-4d8f-4062-bbf2-24340effc837\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.453597 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx5dq\" (UniqueName: \"kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq\") pod \"851c35e9-4d8f-4062-bbf2-24340effc837\" (UID: \"851c35e9-4d8f-4062-bbf2-24340effc837\") " Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.454476 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities" (OuterVolumeSpecName: "utilities") pod "851c35e9-4d8f-4062-bbf2-24340effc837" (UID: "851c35e9-4d8f-4062-bbf2-24340effc837"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.459655 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq" (OuterVolumeSpecName: "kube-api-access-bx5dq") pod "851c35e9-4d8f-4062-bbf2-24340effc837" (UID: "851c35e9-4d8f-4062-bbf2-24340effc837"). InnerVolumeSpecName "kube-api-access-bx5dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.518862 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "851c35e9-4d8f-4062-bbf2-24340effc837" (UID: "851c35e9-4d8f-4062-bbf2-24340effc837"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.558470 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.558542 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/851c35e9-4d8f-4062-bbf2-24340effc837-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.558568 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx5dq\" (UniqueName: \"kubernetes.io/projected/851c35e9-4d8f-4062-bbf2-24340effc837-kube-api-access-bx5dq\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.776102 4867 generic.go:334] "Generic (PLEG): container finished" podID="851c35e9-4d8f-4062-bbf2-24340effc837" containerID="1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006" exitCode=0 Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.776151 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerDied","Data":"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006"} Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.776175 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kddr7" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.776630 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kddr7" event={"ID":"851c35e9-4d8f-4062-bbf2-24340effc837","Type":"ContainerDied","Data":"b0ce08617bc973d617dbe97a980d1ea1aaa6da2433a4cff9be68606ad2872991"} Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.776657 4867 scope.go:117] "RemoveContainer" containerID="1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.814628 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.819305 4867 scope.go:117] "RemoveContainer" containerID="dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.836750 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kddr7"] Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.855456 4867 scope.go:117] "RemoveContainer" containerID="abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.893782 4867 scope.go:117] "RemoveContainer" containerID="1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006" Dec 12 08:46:17 crc kubenswrapper[4867]: E1212 08:46:17.898081 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006\": container with ID starting with 1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006 not found: ID does not exist" containerID="1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.898128 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006"} err="failed to get container status \"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006\": rpc error: code = NotFound desc = could not find container \"1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006\": container with ID starting with 1e4a5d407fb1234e4bc6845014fa3b07b35996e75f51e770a120b7058f183006 not found: ID does not exist" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.898153 4867 scope.go:117] "RemoveContainer" containerID="dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3" Dec 12 08:46:17 crc kubenswrapper[4867]: E1212 08:46:17.899084 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3\": container with ID starting with dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3 not found: ID does not exist" containerID="dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.899142 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3"} err="failed to get container status \"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3\": rpc error: code = NotFound desc = could not find container \"dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3\": container with ID starting with dc761660053bebf1e844ed6abe76391e318ae2f2289bc338f862d2a7de78aba3 not found: ID does not exist" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.899180 4867 scope.go:117] "RemoveContainer" containerID="abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4" Dec 12 08:46:17 crc kubenswrapper[4867]: E1212 08:46:17.899555 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4\": container with ID starting with abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4 not found: ID does not exist" containerID="abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4" Dec 12 08:46:17 crc kubenswrapper[4867]: I1212 08:46:17.899594 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4"} err="failed to get container status \"abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4\": rpc error: code = NotFound desc = could not find container \"abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4\": container with ID starting with abc5ff606cf447991aefda9fa0a82fb278c6c4f7b89d063e46621c86ea6603d4 not found: ID does not exist" Dec 12 08:46:18 crc kubenswrapper[4867]: I1212 08:46:18.855130 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" path="/var/lib/kubelet/pods/851c35e9-4d8f-4062-bbf2-24340effc837/volumes" Dec 12 08:46:48 crc kubenswrapper[4867]: I1212 08:46:48.072110 4867 generic.go:334] "Generic (PLEG): container finished" podID="146d44cb-ac9b-481c-8aa5-0884c7025d47" containerID="f7b123d83d0595f1d274ec35d7b007a111c8ade723144321d6f486ca7847416f" exitCode=0 Dec 12 08:46:48 crc kubenswrapper[4867]: I1212 08:46:48.072209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" event={"ID":"146d44cb-ac9b-481c-8aa5-0884c7025d47","Type":"ContainerDied","Data":"f7b123d83d0595f1d274ec35d7b007a111c8ade723144321d6f486ca7847416f"} Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.538208 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.563044 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key\") pod \"146d44cb-ac9b-481c-8aa5-0884c7025d47\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.563343 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory\") pod \"146d44cb-ac9b-481c-8aa5-0884c7025d47\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.563470 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxk6w\" (UniqueName: \"kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w\") pod \"146d44cb-ac9b-481c-8aa5-0884c7025d47\" (UID: \"146d44cb-ac9b-481c-8aa5-0884c7025d47\") " Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.572118 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w" (OuterVolumeSpecName: "kube-api-access-kxk6w") pod "146d44cb-ac9b-481c-8aa5-0884c7025d47" (UID: "146d44cb-ac9b-481c-8aa5-0884c7025d47"). InnerVolumeSpecName "kube-api-access-kxk6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.605442 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory" (OuterVolumeSpecName: "inventory") pod "146d44cb-ac9b-481c-8aa5-0884c7025d47" (UID: "146d44cb-ac9b-481c-8aa5-0884c7025d47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.607469 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "146d44cb-ac9b-481c-8aa5-0884c7025d47" (UID: "146d44cb-ac9b-481c-8aa5-0884c7025d47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.666152 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.666191 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxk6w\" (UniqueName: \"kubernetes.io/projected/146d44cb-ac9b-481c-8aa5-0884c7025d47-kube-api-access-kxk6w\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:49 crc kubenswrapper[4867]: I1212 08:46:49.666202 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/146d44cb-ac9b-481c-8aa5-0884c7025d47-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.091773 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" event={"ID":"146d44cb-ac9b-481c-8aa5-0884c7025d47","Type":"ContainerDied","Data":"604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137"} Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.092091 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="604cfbefde21ba0bb725cdb4dbcf415efd2dcdb5e46a5be18d025082ebec0137" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.091867 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-lfg7r" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.182403 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vnwc7"] Dec 12 08:46:50 crc kubenswrapper[4867]: E1212 08:46:50.183002 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="registry-server" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.183021 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="registry-server" Dec 12 08:46:50 crc kubenswrapper[4867]: E1212 08:46:50.183038 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="extract-content" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.183047 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="extract-content" Dec 12 08:46:50 crc kubenswrapper[4867]: E1212 08:46:50.183067 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="146d44cb-ac9b-481c-8aa5-0884c7025d47" containerName="download-cache-openstack-openstack-cell1" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.183075 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="146d44cb-ac9b-481c-8aa5-0884c7025d47" containerName="download-cache-openstack-openstack-cell1" Dec 12 08:46:50 crc kubenswrapper[4867]: E1212 08:46:50.183092 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="extract-utilities" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.183099 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="extract-utilities" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.184120 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="146d44cb-ac9b-481c-8aa5-0884c7025d47" containerName="download-cache-openstack-openstack-cell1" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.184145 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="851c35e9-4d8f-4062-bbf2-24340effc837" containerName="registry-server" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.185129 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.188680 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.188745 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.188814 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.189517 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.191045 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vnwc7"] Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.283131 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6l8b\" (UniqueName: \"kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.283257 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.283314 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.386272 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6l8b\" (UniqueName: \"kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.386395 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.386661 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.393278 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.393491 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.408584 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6l8b\" (UniqueName: \"kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b\") pod \"configure-network-openstack-openstack-cell1-vnwc7\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:50 crc kubenswrapper[4867]: I1212 08:46:50.545741 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:46:51 crc kubenswrapper[4867]: I1212 08:46:51.068898 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-vnwc7"] Dec 12 08:46:51 crc kubenswrapper[4867]: W1212 08:46:51.076815 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1220a9e0_08bf_4499_b17a_0e9d53ca3a66.slice/crio-d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b WatchSource:0}: Error finding container d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b: Status 404 returned error can't find the container with id d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b Dec 12 08:46:51 crc kubenswrapper[4867]: I1212 08:46:51.102400 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" event={"ID":"1220a9e0-08bf-4499-b17a-0e9d53ca3a66","Type":"ContainerStarted","Data":"d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b"} Dec 12 08:46:53 crc kubenswrapper[4867]: I1212 08:46:53.125509 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" event={"ID":"1220a9e0-08bf-4499-b17a-0e9d53ca3a66","Type":"ContainerStarted","Data":"bb1fa370542266b2d9215f98f237d539e7240b5a385d9c9fe50f3332c177a812"} Dec 12 08:46:53 crc kubenswrapper[4867]: I1212 08:46:53.147157 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" podStartSLOduration=2.1042629330000002 podStartE2EDuration="3.147138452s" podCreationTimestamp="2025-12-12 08:46:50 +0000 UTC" firstStartedPulling="2025-12-12 08:46:51.079112666 +0000 UTC m=+7098.650493945" lastFinishedPulling="2025-12-12 08:46:52.121988195 +0000 UTC m=+7099.693369464" observedRunningTime="2025-12-12 08:46:53.144389178 +0000 UTC m=+7100.715770467" watchObservedRunningTime="2025-12-12 08:46:53.147138452 +0000 UTC m=+7100.718519721" Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.835949 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.845340 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.860275 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.927303 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr48k\" (UniqueName: \"kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.927723 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:27 crc kubenswrapper[4867]: I1212 08:47:27.927937 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.030277 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr48k\" (UniqueName: \"kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.030326 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.030377 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.030967 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.031033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.052431 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr48k\" (UniqueName: \"kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k\") pod \"community-operators-d7fl6\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.188189 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:28 crc kubenswrapper[4867]: I1212 08:47:28.672887 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:28 crc kubenswrapper[4867]: W1212 08:47:28.685695 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda008dab8_90d2_4b94_8e82_f207cac02eec.slice/crio-d53fd2f4afe99763f56c6887992c4c69c4a8e1aebbb7afc5e5f9284adc963c7d WatchSource:0}: Error finding container d53fd2f4afe99763f56c6887992c4c69c4a8e1aebbb7afc5e5f9284adc963c7d: Status 404 returned error can't find the container with id d53fd2f4afe99763f56c6887992c4c69c4a8e1aebbb7afc5e5f9284adc963c7d Dec 12 08:47:29 crc kubenswrapper[4867]: I1212 08:47:29.461438 4867 generic.go:334] "Generic (PLEG): container finished" podID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerID="d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f" exitCode=0 Dec 12 08:47:29 crc kubenswrapper[4867]: I1212 08:47:29.461514 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerDied","Data":"d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f"} Dec 12 08:47:29 crc kubenswrapper[4867]: I1212 08:47:29.461916 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerStarted","Data":"d53fd2f4afe99763f56c6887992c4c69c4a8e1aebbb7afc5e5f9284adc963c7d"} Dec 12 08:47:29 crc kubenswrapper[4867]: I1212 08:47:29.463552 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:47:30 crc kubenswrapper[4867]: I1212 08:47:30.475545 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerStarted","Data":"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f"} Dec 12 08:47:31 crc kubenswrapper[4867]: I1212 08:47:31.488844 4867 generic.go:334] "Generic (PLEG): container finished" podID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerID="7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f" exitCode=0 Dec 12 08:47:31 crc kubenswrapper[4867]: I1212 08:47:31.488933 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerDied","Data":"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f"} Dec 12 08:47:32 crc kubenswrapper[4867]: I1212 08:47:32.500763 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerStarted","Data":"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9"} Dec 12 08:47:32 crc kubenswrapper[4867]: I1212 08:47:32.526629 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d7fl6" podStartSLOduration=3.020946325 podStartE2EDuration="5.526610115s" podCreationTimestamp="2025-12-12 08:47:27 +0000 UTC" firstStartedPulling="2025-12-12 08:47:29.463275885 +0000 UTC m=+7137.034657164" lastFinishedPulling="2025-12-12 08:47:31.968939685 +0000 UTC m=+7139.540320954" observedRunningTime="2025-12-12 08:47:32.520637756 +0000 UTC m=+7140.092019025" watchObservedRunningTime="2025-12-12 08:47:32.526610115 +0000 UTC m=+7140.097991384" Dec 12 08:47:38 crc kubenswrapper[4867]: I1212 08:47:38.190314 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:38 crc kubenswrapper[4867]: I1212 08:47:38.190961 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:38 crc kubenswrapper[4867]: I1212 08:47:38.256599 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:38 crc kubenswrapper[4867]: I1212 08:47:38.621676 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:38 crc kubenswrapper[4867]: I1212 08:47:38.668956 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:40 crc kubenswrapper[4867]: I1212 08:47:40.571728 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d7fl6" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="registry-server" containerID="cri-o://a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9" gracePeriod=2 Dec 12 08:47:40 crc kubenswrapper[4867]: I1212 08:47:40.994282 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.108544 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr48k\" (UniqueName: \"kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k\") pod \"a008dab8-90d2-4b94-8e82-f207cac02eec\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.108769 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities\") pod \"a008dab8-90d2-4b94-8e82-f207cac02eec\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.108840 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content\") pod \"a008dab8-90d2-4b94-8e82-f207cac02eec\" (UID: \"a008dab8-90d2-4b94-8e82-f207cac02eec\") " Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.109557 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities" (OuterVolumeSpecName: "utilities") pod "a008dab8-90d2-4b94-8e82-f207cac02eec" (UID: "a008dab8-90d2-4b94-8e82-f207cac02eec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.109807 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.113564 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k" (OuterVolumeSpecName: "kube-api-access-rr48k") pod "a008dab8-90d2-4b94-8e82-f207cac02eec" (UID: "a008dab8-90d2-4b94-8e82-f207cac02eec"). InnerVolumeSpecName "kube-api-access-rr48k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.154006 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a008dab8-90d2-4b94-8e82-f207cac02eec" (UID: "a008dab8-90d2-4b94-8e82-f207cac02eec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.211656 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr48k\" (UniqueName: \"kubernetes.io/projected/a008dab8-90d2-4b94-8e82-f207cac02eec-kube-api-access-rr48k\") on node \"crc\" DevicePath \"\"" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.211694 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a008dab8-90d2-4b94-8e82-f207cac02eec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.588653 4867 generic.go:334] "Generic (PLEG): container finished" podID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerID="a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9" exitCode=0 Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.588707 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerDied","Data":"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9"} Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.588738 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7fl6" event={"ID":"a008dab8-90d2-4b94-8e82-f207cac02eec","Type":"ContainerDied","Data":"d53fd2f4afe99763f56c6887992c4c69c4a8e1aebbb7afc5e5f9284adc963c7d"} Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.588759 4867 scope.go:117] "RemoveContainer" containerID="a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.588951 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7fl6" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.627715 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.629408 4867 scope.go:117] "RemoveContainer" containerID="7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.641076 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d7fl6"] Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.659283 4867 scope.go:117] "RemoveContainer" containerID="d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.696373 4867 scope.go:117] "RemoveContainer" containerID="a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9" Dec 12 08:47:41 crc kubenswrapper[4867]: E1212 08:47:41.696827 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9\": container with ID starting with a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9 not found: ID does not exist" containerID="a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.696870 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9"} err="failed to get container status \"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9\": rpc error: code = NotFound desc = could not find container \"a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9\": container with ID starting with a229496e846fd439eb681602d7795562b378390ee21aee836256afc0158a8db9 not found: ID does not exist" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.696894 4867 scope.go:117] "RemoveContainer" containerID="7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f" Dec 12 08:47:41 crc kubenswrapper[4867]: E1212 08:47:41.697191 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f\": container with ID starting with 7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f not found: ID does not exist" containerID="7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.697218 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f"} err="failed to get container status \"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f\": rpc error: code = NotFound desc = could not find container \"7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f\": container with ID starting with 7b6f18ae16c4f23a7bb68f7d32badd909241729f6e2c2e753137277452d8609f not found: ID does not exist" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.697256 4867 scope.go:117] "RemoveContainer" containerID="d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f" Dec 12 08:47:41 crc kubenswrapper[4867]: E1212 08:47:41.697874 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f\": container with ID starting with d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f not found: ID does not exist" containerID="d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f" Dec 12 08:47:41 crc kubenswrapper[4867]: I1212 08:47:41.697916 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f"} err="failed to get container status \"d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f\": rpc error: code = NotFound desc = could not find container \"d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f\": container with ID starting with d733d0eb5057b52ee53c0aa9250a1be04a3f31f31b5a740d78c4df630497751f not found: ID does not exist" Dec 12 08:47:42 crc kubenswrapper[4867]: I1212 08:47:42.851287 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" path="/var/lib/kubelet/pods/a008dab8-90d2-4b94-8e82-f207cac02eec/volumes" Dec 12 08:47:58 crc kubenswrapper[4867]: I1212 08:47:58.989472 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:47:58 crc kubenswrapper[4867]: I1212 08:47:58.990681 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:48:13 crc kubenswrapper[4867]: I1212 08:48:13.900559 4867 generic.go:334] "Generic (PLEG): container finished" podID="1220a9e0-08bf-4499-b17a-0e9d53ca3a66" containerID="bb1fa370542266b2d9215f98f237d539e7240b5a385d9c9fe50f3332c177a812" exitCode=0 Dec 12 08:48:13 crc kubenswrapper[4867]: I1212 08:48:13.900686 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" event={"ID":"1220a9e0-08bf-4499-b17a-0e9d53ca3a66","Type":"ContainerDied","Data":"bb1fa370542266b2d9215f98f237d539e7240b5a385d9c9fe50f3332c177a812"} Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.315383 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.417656 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key\") pod \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.417971 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6l8b\" (UniqueName: \"kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b\") pod \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.418439 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory\") pod \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\" (UID: \"1220a9e0-08bf-4499-b17a-0e9d53ca3a66\") " Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.423449 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b" (OuterVolumeSpecName: "kube-api-access-f6l8b") pod "1220a9e0-08bf-4499-b17a-0e9d53ca3a66" (UID: "1220a9e0-08bf-4499-b17a-0e9d53ca3a66"). InnerVolumeSpecName "kube-api-access-f6l8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.445589 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1220a9e0-08bf-4499-b17a-0e9d53ca3a66" (UID: "1220a9e0-08bf-4499-b17a-0e9d53ca3a66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.449939 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory" (OuterVolumeSpecName: "inventory") pod "1220a9e0-08bf-4499-b17a-0e9d53ca3a66" (UID: "1220a9e0-08bf-4499-b17a-0e9d53ca3a66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.521633 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6l8b\" (UniqueName: \"kubernetes.io/projected/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-kube-api-access-f6l8b\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.521677 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.521691 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1220a9e0-08bf-4499-b17a-0e9d53ca3a66-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.925323 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" event={"ID":"1220a9e0-08bf-4499-b17a-0e9d53ca3a66","Type":"ContainerDied","Data":"d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b"} Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.925366 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c8ef7cb3f8c7fdd909c5294ab72d60d3f4bd94cc46c45d1ee60de3f041cc1b" Dec 12 08:48:15 crc kubenswrapper[4867]: I1212 08:48:15.925430 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-vnwc7" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.030329 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-w886f"] Dec 12 08:48:16 crc kubenswrapper[4867]: E1212 08:48:16.050065 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="registry-server" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.050105 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="registry-server" Dec 12 08:48:16 crc kubenswrapper[4867]: E1212 08:48:16.050161 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="extract-utilities" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.050176 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="extract-utilities" Dec 12 08:48:16 crc kubenswrapper[4867]: E1212 08:48:16.050203 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1220a9e0-08bf-4499-b17a-0e9d53ca3a66" containerName="configure-network-openstack-openstack-cell1" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.050249 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1220a9e0-08bf-4499-b17a-0e9d53ca3a66" containerName="configure-network-openstack-openstack-cell1" Dec 12 08:48:16 crc kubenswrapper[4867]: E1212 08:48:16.050282 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="extract-content" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.050291 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="extract-content" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.050764 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a008dab8-90d2-4b94-8e82-f207cac02eec" containerName="registry-server" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.051024 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1220a9e0-08bf-4499-b17a-0e9d53ca3a66" containerName="configure-network-openstack-openstack-cell1" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.052768 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.055718 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.056264 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.056415 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.056817 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.060866 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-w886f"] Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.154505 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmqm4\" (UniqueName: \"kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.154592 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.154788 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.256873 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmqm4\" (UniqueName: \"kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.256970 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.257035 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.268814 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.273210 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.274907 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmqm4\" (UniqueName: \"kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4\") pod \"validate-network-openstack-openstack-cell1-w886f\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.374942 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:16 crc kubenswrapper[4867]: I1212 08:48:16.934616 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-w886f"] Dec 12 08:48:17 crc kubenswrapper[4867]: I1212 08:48:17.946808 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-w886f" event={"ID":"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc","Type":"ContainerStarted","Data":"4633fa862bb40ca39a17428f3075539334b8a9641e03869b460d177a97694efd"} Dec 12 08:48:17 crc kubenswrapper[4867]: I1212 08:48:17.948360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-w886f" event={"ID":"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc","Type":"ContainerStarted","Data":"8df4a6c934c2589596e41bf1082f0b8a411f0dd8793d5fae012259a7e8b5bdf0"} Dec 12 08:48:17 crc kubenswrapper[4867]: I1212 08:48:17.978072 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-w886f" podStartSLOduration=2.503749916 podStartE2EDuration="2.977966589s" podCreationTimestamp="2025-12-12 08:48:15 +0000 UTC" firstStartedPulling="2025-12-12 08:48:16.952132637 +0000 UTC m=+7184.523513916" lastFinishedPulling="2025-12-12 08:48:17.42634932 +0000 UTC m=+7184.997730589" observedRunningTime="2025-12-12 08:48:17.965003089 +0000 UTC m=+7185.536384398" watchObservedRunningTime="2025-12-12 08:48:17.977966589 +0000 UTC m=+7185.549347868" Dec 12 08:48:23 crc kubenswrapper[4867]: I1212 08:48:23.011430 4867 generic.go:334] "Generic (PLEG): container finished" podID="bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" containerID="4633fa862bb40ca39a17428f3075539334b8a9641e03869b460d177a97694efd" exitCode=0 Dec 12 08:48:23 crc kubenswrapper[4867]: I1212 08:48:23.011506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-w886f" event={"ID":"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc","Type":"ContainerDied","Data":"4633fa862bb40ca39a17428f3075539334b8a9641e03869b460d177a97694efd"} Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.460475 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.631221 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory\") pod \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.631325 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmqm4\" (UniqueName: \"kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4\") pod \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.631368 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key\") pod \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\" (UID: \"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc\") " Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.639532 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4" (OuterVolumeSpecName: "kube-api-access-qmqm4") pod "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" (UID: "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc"). InnerVolumeSpecName "kube-api-access-qmqm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.666939 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory" (OuterVolumeSpecName: "inventory") pod "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" (UID: "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.673842 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" (UID: "bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.733458 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.733606 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmqm4\" (UniqueName: \"kubernetes.io/projected/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-kube-api-access-qmqm4\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:24 crc kubenswrapper[4867]: I1212 08:48:24.733687 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.034768 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-w886f" event={"ID":"bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc","Type":"ContainerDied","Data":"8df4a6c934c2589596e41bf1082f0b8a411f0dd8793d5fae012259a7e8b5bdf0"} Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.034813 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8df4a6c934c2589596e41bf1082f0b8a411f0dd8793d5fae012259a7e8b5bdf0" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.034883 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-w886f" Dec 12 08:48:25 crc kubenswrapper[4867]: E1212 08:48:25.042179 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd4446dc_714b_4d9f_81f3_c1dcdeb7b8bc.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd4446dc_714b_4d9f_81f3_c1dcdeb7b8bc.slice/crio-8df4a6c934c2589596e41bf1082f0b8a411f0dd8793d5fae012259a7e8b5bdf0\": RecentStats: unable to find data in memory cache]" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.108303 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-99v58"] Dec 12 08:48:25 crc kubenswrapper[4867]: E1212 08:48:25.108954 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" containerName="validate-network-openstack-openstack-cell1" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.108978 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" containerName="validate-network-openstack-openstack-cell1" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.109215 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc" containerName="validate-network-openstack-openstack-cell1" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.110191 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.115198 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.115430 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.115578 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.116053 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.130126 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-99v58"] Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.240973 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.241767 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxp8d\" (UniqueName: \"kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.241844 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.344550 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.344619 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxp8d\" (UniqueName: \"kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.344691 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.348391 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.349111 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.369475 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxp8d\" (UniqueName: \"kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d\") pod \"install-os-openstack-openstack-cell1-99v58\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.427807 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:48:25 crc kubenswrapper[4867]: I1212 08:48:25.976841 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-99v58"] Dec 12 08:48:26 crc kubenswrapper[4867]: I1212 08:48:26.043914 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-99v58" event={"ID":"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8","Type":"ContainerStarted","Data":"bc710c4bfe6b0c5ab8f29d7fadc76cdf3c743cdaa44aa06139c013c6f4a4e8fd"} Dec 12 08:48:28 crc kubenswrapper[4867]: I1212 08:48:28.062871 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-99v58" event={"ID":"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8","Type":"ContainerStarted","Data":"ece8e5b4626225cdf5d8496778afa78c9f68662fac7bcd3d3a8b1848f9ec0206"} Dec 12 08:48:28 crc kubenswrapper[4867]: I1212 08:48:28.078653 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-99v58" podStartSLOduration=1.403714176 podStartE2EDuration="3.078633019s" podCreationTimestamp="2025-12-12 08:48:25 +0000 UTC" firstStartedPulling="2025-12-12 08:48:25.991321857 +0000 UTC m=+7193.562703126" lastFinishedPulling="2025-12-12 08:48:27.6662407 +0000 UTC m=+7195.237621969" observedRunningTime="2025-12-12 08:48:28.078201619 +0000 UTC m=+7195.649582888" watchObservedRunningTime="2025-12-12 08:48:28.078633019 +0000 UTC m=+7195.650014288" Dec 12 08:48:28 crc kubenswrapper[4867]: I1212 08:48:28.989509 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:48:28 crc kubenswrapper[4867]: I1212 08:48:28.989569 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:48:58 crc kubenswrapper[4867]: I1212 08:48:58.988854 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:48:58 crc kubenswrapper[4867]: I1212 08:48:58.989860 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:48:58 crc kubenswrapper[4867]: I1212 08:48:58.989935 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:48:58 crc kubenswrapper[4867]: I1212 08:48:58.991149 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:48:58 crc kubenswrapper[4867]: I1212 08:48:58.991283 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2" gracePeriod=600 Dec 12 08:48:59 crc kubenswrapper[4867]: I1212 08:48:59.410865 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2" exitCode=0 Dec 12 08:48:59 crc kubenswrapper[4867]: I1212 08:48:59.410909 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2"} Dec 12 08:48:59 crc kubenswrapper[4867]: I1212 08:48:59.411264 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8"} Dec 12 08:48:59 crc kubenswrapper[4867]: I1212 08:48:59.411289 4867 scope.go:117] "RemoveContainer" containerID="fc54b9f7c6c3e427d16782ec27151be45479ebd3b7c6b07b42028eed5acb1a35" Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.877384 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.880084 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.912935 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.960930 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.961004 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:04 crc kubenswrapper[4867]: I1212 08:49:04.961082 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkllv\" (UniqueName: \"kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.063209 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.063294 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.063356 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkllv\" (UniqueName: \"kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.063872 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.063872 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.083015 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkllv\" (UniqueName: \"kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv\") pod \"redhat-operators-bv6zw\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.204927 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:05 crc kubenswrapper[4867]: I1212 08:49:05.651707 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:06 crc kubenswrapper[4867]: E1212 08:49:06.088402 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38b18806_be93_4c6b_946b_a9385bc9c3cf.slice/crio-f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:49:06 crc kubenswrapper[4867]: I1212 08:49:06.477081 4867 generic.go:334] "Generic (PLEG): container finished" podID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerID="f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2" exitCode=0 Dec 12 08:49:06 crc kubenswrapper[4867]: I1212 08:49:06.477137 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerDied","Data":"f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2"} Dec 12 08:49:06 crc kubenswrapper[4867]: I1212 08:49:06.477193 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerStarted","Data":"d261aac92f6b033019e537b8bf5b466842ebe7f187eb5e6ccb1d36ed386f6cb1"} Dec 12 08:49:07 crc kubenswrapper[4867]: I1212 08:49:07.489151 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerStarted","Data":"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0"} Dec 12 08:49:08 crc kubenswrapper[4867]: I1212 08:49:08.499244 4867 generic.go:334] "Generic (PLEG): container finished" podID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerID="73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0" exitCode=0 Dec 12 08:49:08 crc kubenswrapper[4867]: I1212 08:49:08.499301 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerDied","Data":"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0"} Dec 12 08:49:10 crc kubenswrapper[4867]: I1212 08:49:10.521346 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerStarted","Data":"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930"} Dec 12 08:49:10 crc kubenswrapper[4867]: I1212 08:49:10.543996 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bv6zw" podStartSLOduration=3.677474943 podStartE2EDuration="6.543974705s" podCreationTimestamp="2025-12-12 08:49:04 +0000 UTC" firstStartedPulling="2025-12-12 08:49:06.480417387 +0000 UTC m=+7234.051798656" lastFinishedPulling="2025-12-12 08:49:09.346917149 +0000 UTC m=+7236.918298418" observedRunningTime="2025-12-12 08:49:10.540845193 +0000 UTC m=+7238.112226462" watchObservedRunningTime="2025-12-12 08:49:10.543974705 +0000 UTC m=+7238.115355974" Dec 12 08:49:13 crc kubenswrapper[4867]: I1212 08:49:13.557068 4867 generic.go:334] "Generic (PLEG): container finished" podID="a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" containerID="ece8e5b4626225cdf5d8496778afa78c9f68662fac7bcd3d3a8b1848f9ec0206" exitCode=0 Dec 12 08:49:13 crc kubenswrapper[4867]: I1212 08:49:13.557152 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-99v58" event={"ID":"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8","Type":"ContainerDied","Data":"ece8e5b4626225cdf5d8496778afa78c9f68662fac7bcd3d3a8b1848f9ec0206"} Dec 12 08:49:14 crc kubenswrapper[4867]: I1212 08:49:14.965363 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.083411 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory\") pod \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.083477 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key\") pod \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.083697 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxp8d\" (UniqueName: \"kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d\") pod \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\" (UID: \"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8\") " Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.089786 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d" (OuterVolumeSpecName: "kube-api-access-jxp8d") pod "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" (UID: "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8"). InnerVolumeSpecName "kube-api-access-jxp8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.112634 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory" (OuterVolumeSpecName: "inventory") pod "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" (UID: "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.119421 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" (UID: "a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.186182 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.186375 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.186452 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxp8d\" (UniqueName: \"kubernetes.io/projected/a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8-kube-api-access-jxp8d\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.205386 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.205457 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.250375 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.574593 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-99v58" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.574732 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-99v58" event={"ID":"a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8","Type":"ContainerDied","Data":"bc710c4bfe6b0c5ab8f29d7fadc76cdf3c743cdaa44aa06139c013c6f4a4e8fd"} Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.574780 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc710c4bfe6b0c5ab8f29d7fadc76cdf3c743cdaa44aa06139c013c6f4a4e8fd" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.630729 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.670769 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s6wxg"] Dec 12 08:49:15 crc kubenswrapper[4867]: E1212 08:49:15.678213 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" containerName="install-os-openstack-openstack-cell1" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.678305 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" containerName="install-os-openstack-openstack-cell1" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.680230 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8" containerName="install-os-openstack-openstack-cell1" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.681781 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.687514 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.688517 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.689446 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.689712 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.706578 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s6wxg"] Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.800488 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.800749 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.801364 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdvgf\" (UniqueName: \"kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.903282 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.903493 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdvgf\" (UniqueName: \"kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.903539 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.907805 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.907907 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:15 crc kubenswrapper[4867]: I1212 08:49:15.919227 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdvgf\" (UniqueName: \"kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf\") pod \"configure-os-openstack-openstack-cell1-s6wxg\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:16 crc kubenswrapper[4867]: I1212 08:49:16.005777 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:49:16 crc kubenswrapper[4867]: I1212 08:49:16.509810 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s6wxg"] Dec 12 08:49:16 crc kubenswrapper[4867]: I1212 08:49:16.587503 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" event={"ID":"88b38f78-a939-4410-8b4b-a584d7a12f34","Type":"ContainerStarted","Data":"86b6fa39f872ac4d1e12a39a512dd528fb400b67603fc735ff1bd8a5461eb383"} Dec 12 08:49:19 crc kubenswrapper[4867]: I1212 08:49:19.615338 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" event={"ID":"88b38f78-a939-4410-8b4b-a584d7a12f34","Type":"ContainerStarted","Data":"6b3ff6d7ccfbfe13ec2888e49ff382829ff8a36e7f8252aa8e1a3988841ca00e"} Dec 12 08:49:19 crc kubenswrapper[4867]: I1212 08:49:19.643936 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" podStartSLOduration=2.751548156 podStartE2EDuration="4.643909584s" podCreationTimestamp="2025-12-12 08:49:15 +0000 UTC" firstStartedPulling="2025-12-12 08:49:16.512462565 +0000 UTC m=+7244.083843834" lastFinishedPulling="2025-12-12 08:49:18.404823993 +0000 UTC m=+7245.976205262" observedRunningTime="2025-12-12 08:49:19.634643789 +0000 UTC m=+7247.206025058" watchObservedRunningTime="2025-12-12 08:49:19.643909584 +0000 UTC m=+7247.215290863" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.049966 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.050584 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bv6zw" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="registry-server" containerID="cri-o://df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930" gracePeriod=2 Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.507254 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.598932 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkllv\" (UniqueName: \"kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv\") pod \"38b18806-be93-4c6b-946b-a9385bc9c3cf\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.598993 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content\") pod \"38b18806-be93-4c6b-946b-a9385bc9c3cf\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.599065 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities\") pod \"38b18806-be93-4c6b-946b-a9385bc9c3cf\" (UID: \"38b18806-be93-4c6b-946b-a9385bc9c3cf\") " Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.599988 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities" (OuterVolumeSpecName: "utilities") pod "38b18806-be93-4c6b-946b-a9385bc9c3cf" (UID: "38b18806-be93-4c6b-946b-a9385bc9c3cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.604378 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv" (OuterVolumeSpecName: "kube-api-access-kkllv") pod "38b18806-be93-4c6b-946b-a9385bc9c3cf" (UID: "38b18806-be93-4c6b-946b-a9385bc9c3cf"). InnerVolumeSpecName "kube-api-access-kkllv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.627192 4867 generic.go:334] "Generic (PLEG): container finished" podID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerID="df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930" exitCode=0 Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.627274 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerDied","Data":"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930"} Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.627333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bv6zw" event={"ID":"38b18806-be93-4c6b-946b-a9385bc9c3cf","Type":"ContainerDied","Data":"d261aac92f6b033019e537b8bf5b466842ebe7f187eb5e6ccb1d36ed386f6cb1"} Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.627360 4867 scope.go:117] "RemoveContainer" containerID="df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.627287 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bv6zw" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.673582 4867 scope.go:117] "RemoveContainer" containerID="73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.704374 4867 scope.go:117] "RemoveContainer" containerID="f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.705446 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkllv\" (UniqueName: \"kubernetes.io/projected/38b18806-be93-4c6b-946b-a9385bc9c3cf-kube-api-access-kkllv\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.705476 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.729414 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38b18806-be93-4c6b-946b-a9385bc9c3cf" (UID: "38b18806-be93-4c6b-946b-a9385bc9c3cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.748919 4867 scope.go:117] "RemoveContainer" containerID="df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930" Dec 12 08:49:20 crc kubenswrapper[4867]: E1212 08:49:20.749477 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930\": container with ID starting with df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930 not found: ID does not exist" containerID="df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.749520 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930"} err="failed to get container status \"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930\": rpc error: code = NotFound desc = could not find container \"df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930\": container with ID starting with df242fc5d5a095172a69035a6a14d7b78f391c55b6324941209a58d36b814930 not found: ID does not exist" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.749545 4867 scope.go:117] "RemoveContainer" containerID="73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0" Dec 12 08:49:20 crc kubenswrapper[4867]: E1212 08:49:20.749854 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0\": container with ID starting with 73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0 not found: ID does not exist" containerID="73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.749895 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0"} err="failed to get container status \"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0\": rpc error: code = NotFound desc = could not find container \"73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0\": container with ID starting with 73c34af82f3a749bcb9e077816f44639986dd293eb501edbfa5eb47c081584d0 not found: ID does not exist" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.749923 4867 scope.go:117] "RemoveContainer" containerID="f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2" Dec 12 08:49:20 crc kubenswrapper[4867]: E1212 08:49:20.750385 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2\": container with ID starting with f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2 not found: ID does not exist" containerID="f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.750422 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2"} err="failed to get container status \"f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2\": rpc error: code = NotFound desc = could not find container \"f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2\": container with ID starting with f568576a7f4e1475386d65f86306a0bfbd7d6274996a5ac91fb75afae273aaa2 not found: ID does not exist" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.807146 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38b18806-be93-4c6b-946b-a9385bc9c3cf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.948122 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:20 crc kubenswrapper[4867]: I1212 08:49:20.955428 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bv6zw"] Dec 12 08:49:22 crc kubenswrapper[4867]: I1212 08:49:22.851320 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" path="/var/lib/kubelet/pods/38b18806-be93-4c6b-946b-a9385bc9c3cf/volumes" Dec 12 08:50:04 crc kubenswrapper[4867]: I1212 08:50:04.017938 4867 generic.go:334] "Generic (PLEG): container finished" podID="88b38f78-a939-4410-8b4b-a584d7a12f34" containerID="6b3ff6d7ccfbfe13ec2888e49ff382829ff8a36e7f8252aa8e1a3988841ca00e" exitCode=0 Dec 12 08:50:04 crc kubenswrapper[4867]: I1212 08:50:04.018044 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" event={"ID":"88b38f78-a939-4410-8b4b-a584d7a12f34","Type":"ContainerDied","Data":"6b3ff6d7ccfbfe13ec2888e49ff382829ff8a36e7f8252aa8e1a3988841ca00e"} Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.567595 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.688856 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdvgf\" (UniqueName: \"kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf\") pod \"88b38f78-a939-4410-8b4b-a584d7a12f34\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.688992 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key\") pod \"88b38f78-a939-4410-8b4b-a584d7a12f34\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.689059 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory\") pod \"88b38f78-a939-4410-8b4b-a584d7a12f34\" (UID: \"88b38f78-a939-4410-8b4b-a584d7a12f34\") " Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.695125 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf" (OuterVolumeSpecName: "kube-api-access-mdvgf") pod "88b38f78-a939-4410-8b4b-a584d7a12f34" (UID: "88b38f78-a939-4410-8b4b-a584d7a12f34"). InnerVolumeSpecName "kube-api-access-mdvgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.719527 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88b38f78-a939-4410-8b4b-a584d7a12f34" (UID: "88b38f78-a939-4410-8b4b-a584d7a12f34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.721902 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory" (OuterVolumeSpecName: "inventory") pod "88b38f78-a939-4410-8b4b-a584d7a12f34" (UID: "88b38f78-a939-4410-8b4b-a584d7a12f34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.791314 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.791503 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88b38f78-a939-4410-8b4b-a584d7a12f34-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:05 crc kubenswrapper[4867]: I1212 08:50:05.791806 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdvgf\" (UniqueName: \"kubernetes.io/projected/88b38f78-a939-4410-8b4b-a584d7a12f34-kube-api-access-mdvgf\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.075011 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" event={"ID":"88b38f78-a939-4410-8b4b-a584d7a12f34","Type":"ContainerDied","Data":"86b6fa39f872ac4d1e12a39a512dd528fb400b67603fc735ff1bd8a5461eb383"} Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.075380 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86b6fa39f872ac4d1e12a39a512dd528fb400b67603fc735ff1bd8a5461eb383" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.075129 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s6wxg" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.161491 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-p46gb"] Dec 12 08:50:06 crc kubenswrapper[4867]: E1212 08:50:06.161994 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b38f78-a939-4410-8b4b-a584d7a12f34" containerName="configure-os-openstack-openstack-cell1" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162011 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b38f78-a939-4410-8b4b-a584d7a12f34" containerName="configure-os-openstack-openstack-cell1" Dec 12 08:50:06 crc kubenswrapper[4867]: E1212 08:50:06.162037 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="extract-content" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162046 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="extract-content" Dec 12 08:50:06 crc kubenswrapper[4867]: E1212 08:50:06.162063 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="extract-utilities" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162071 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="extract-utilities" Dec 12 08:50:06 crc kubenswrapper[4867]: E1212 08:50:06.162094 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="registry-server" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162101 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="registry-server" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162373 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="38b18806-be93-4c6b-946b-a9385bc9c3cf" containerName="registry-server" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.162392 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b38f78-a939-4410-8b4b-a584d7a12f34" containerName="configure-os-openstack-openstack-cell1" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.163304 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.166425 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.166578 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.170333 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.170530 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.171940 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-p46gb"] Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.300557 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.300611 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctjlp\" (UniqueName: \"kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.300743 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.402623 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.402888 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctjlp\" (UniqueName: \"kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.403129 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.407191 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.409800 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.419734 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctjlp\" (UniqueName: \"kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp\") pod \"ssh-known-hosts-openstack-p46gb\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:06 crc kubenswrapper[4867]: I1212 08:50:06.492077 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:07 crc kubenswrapper[4867]: I1212 08:50:07.086322 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-p46gb"] Dec 12 08:50:08 crc kubenswrapper[4867]: I1212 08:50:08.095061 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-p46gb" event={"ID":"f78f4a6a-78a2-4f00-9835-768392104c41","Type":"ContainerStarted","Data":"4de74c2082a3ab7ef9a59a87b2f932fa0b9305e8b06a5609da526f2b5d42006f"} Dec 12 08:50:08 crc kubenswrapper[4867]: I1212 08:50:08.095613 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-p46gb" event={"ID":"f78f4a6a-78a2-4f00-9835-768392104c41","Type":"ContainerStarted","Data":"b758874d9f1bc6e5b00744831c2da0c89c323f102546f3f2ce876871498e2982"} Dec 12 08:50:08 crc kubenswrapper[4867]: I1212 08:50:08.118252 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-p46gb" podStartSLOduration=1.482170645 podStartE2EDuration="2.118217313s" podCreationTimestamp="2025-12-12 08:50:06 +0000 UTC" firstStartedPulling="2025-12-12 08:50:07.09320645 +0000 UTC m=+7294.664587719" lastFinishedPulling="2025-12-12 08:50:07.729253118 +0000 UTC m=+7295.300634387" observedRunningTime="2025-12-12 08:50:08.109360088 +0000 UTC m=+7295.680741367" watchObservedRunningTime="2025-12-12 08:50:08.118217313 +0000 UTC m=+7295.689598582" Dec 12 08:50:17 crc kubenswrapper[4867]: I1212 08:50:17.189048 4867 generic.go:334] "Generic (PLEG): container finished" podID="f78f4a6a-78a2-4f00-9835-768392104c41" containerID="4de74c2082a3ab7ef9a59a87b2f932fa0b9305e8b06a5609da526f2b5d42006f" exitCode=0 Dec 12 08:50:17 crc kubenswrapper[4867]: I1212 08:50:17.189155 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-p46gb" event={"ID":"f78f4a6a-78a2-4f00-9835-768392104c41","Type":"ContainerDied","Data":"4de74c2082a3ab7ef9a59a87b2f932fa0b9305e8b06a5609da526f2b5d42006f"} Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.606060 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.772428 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0\") pod \"f78f4a6a-78a2-4f00-9835-768392104c41\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.772563 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1\") pod \"f78f4a6a-78a2-4f00-9835-768392104c41\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.772629 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctjlp\" (UniqueName: \"kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp\") pod \"f78f4a6a-78a2-4f00-9835-768392104c41\" (UID: \"f78f4a6a-78a2-4f00-9835-768392104c41\") " Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.786369 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp" (OuterVolumeSpecName: "kube-api-access-ctjlp") pod "f78f4a6a-78a2-4f00-9835-768392104c41" (UID: "f78f4a6a-78a2-4f00-9835-768392104c41"). InnerVolumeSpecName "kube-api-access-ctjlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.803416 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "f78f4a6a-78a2-4f00-9835-768392104c41" (UID: "f78f4a6a-78a2-4f00-9835-768392104c41"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.807369 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "f78f4a6a-78a2-4f00-9835-768392104c41" (UID: "f78f4a6a-78a2-4f00-9835-768392104c41"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.878094 4867 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.878129 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/f78f4a6a-78a2-4f00-9835-768392104c41-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:18 crc kubenswrapper[4867]: I1212 08:50:18.878141 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctjlp\" (UniqueName: \"kubernetes.io/projected/f78f4a6a-78a2-4f00-9835-768392104c41-kube-api-access-ctjlp\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.212111 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-p46gb" event={"ID":"f78f4a6a-78a2-4f00-9835-768392104c41","Type":"ContainerDied","Data":"b758874d9f1bc6e5b00744831c2da0c89c323f102546f3f2ce876871498e2982"} Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.212152 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b758874d9f1bc6e5b00744831c2da0c89c323f102546f3f2ce876871498e2982" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.212194 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-p46gb" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.350594 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-76bjl"] Dec 12 08:50:19 crc kubenswrapper[4867]: E1212 08:50:19.351099 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78f4a6a-78a2-4f00-9835-768392104c41" containerName="ssh-known-hosts-openstack" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.351113 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78f4a6a-78a2-4f00-9835-768392104c41" containerName="ssh-known-hosts-openstack" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.351386 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78f4a6a-78a2-4f00-9835-768392104c41" containerName="ssh-known-hosts-openstack" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.352114 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.354757 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.354940 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.355146 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.357707 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.371345 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-76bjl"] Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.491070 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njcc7\" (UniqueName: \"kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.491582 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.491797 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.593894 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.594084 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njcc7\" (UniqueName: \"kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.594266 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.598917 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.604772 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.623895 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njcc7\" (UniqueName: \"kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7\") pod \"run-os-openstack-openstack-cell1-76bjl\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:19 crc kubenswrapper[4867]: I1212 08:50:19.675033 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:20 crc kubenswrapper[4867]: I1212 08:50:20.368417 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-76bjl"] Dec 12 08:50:20 crc kubenswrapper[4867]: W1212 08:50:20.370606 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6760547_a35d_4676_9ef9_ca5b94a76fb2.slice/crio-92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530 WatchSource:0}: Error finding container 92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530: Status 404 returned error can't find the container with id 92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530 Dec 12 08:50:21 crc kubenswrapper[4867]: I1212 08:50:21.227754 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-76bjl" event={"ID":"a6760547-a35d-4676-9ef9-ca5b94a76fb2","Type":"ContainerStarted","Data":"d326d05dec36c828a1e614ecccf2e77ab102bdfca0f314e8623fb376c6786485"} Dec 12 08:50:21 crc kubenswrapper[4867]: I1212 08:50:21.227810 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-76bjl" event={"ID":"a6760547-a35d-4676-9ef9-ca5b94a76fb2","Type":"ContainerStarted","Data":"92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530"} Dec 12 08:50:21 crc kubenswrapper[4867]: I1212 08:50:21.248251 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-76bjl" podStartSLOduration=1.7241499120000001 podStartE2EDuration="2.248197732s" podCreationTimestamp="2025-12-12 08:50:19 +0000 UTC" firstStartedPulling="2025-12-12 08:50:20.372975764 +0000 UTC m=+7307.944357033" lastFinishedPulling="2025-12-12 08:50:20.897023574 +0000 UTC m=+7308.468404853" observedRunningTime="2025-12-12 08:50:21.241041826 +0000 UTC m=+7308.812423095" watchObservedRunningTime="2025-12-12 08:50:21.248197732 +0000 UTC m=+7308.819579001" Dec 12 08:50:29 crc kubenswrapper[4867]: I1212 08:50:29.298135 4867 generic.go:334] "Generic (PLEG): container finished" podID="a6760547-a35d-4676-9ef9-ca5b94a76fb2" containerID="d326d05dec36c828a1e614ecccf2e77ab102bdfca0f314e8623fb376c6786485" exitCode=0 Dec 12 08:50:29 crc kubenswrapper[4867]: I1212 08:50:29.298195 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-76bjl" event={"ID":"a6760547-a35d-4676-9ef9-ca5b94a76fb2","Type":"ContainerDied","Data":"d326d05dec36c828a1e614ecccf2e77ab102bdfca0f314e8623fb376c6786485"} Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.740534 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.934590 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key\") pod \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.934938 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njcc7\" (UniqueName: \"kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7\") pod \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.935030 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory\") pod \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\" (UID: \"a6760547-a35d-4676-9ef9-ca5b94a76fb2\") " Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.940331 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7" (OuterVolumeSpecName: "kube-api-access-njcc7") pod "a6760547-a35d-4676-9ef9-ca5b94a76fb2" (UID: "a6760547-a35d-4676-9ef9-ca5b94a76fb2"). InnerVolumeSpecName "kube-api-access-njcc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.966701 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory" (OuterVolumeSpecName: "inventory") pod "a6760547-a35d-4676-9ef9-ca5b94a76fb2" (UID: "a6760547-a35d-4676-9ef9-ca5b94a76fb2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:30 crc kubenswrapper[4867]: I1212 08:50:30.992788 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6760547-a35d-4676-9ef9-ca5b94a76fb2" (UID: "a6760547-a35d-4676-9ef9-ca5b94a76fb2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.040497 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.040530 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njcc7\" (UniqueName: \"kubernetes.io/projected/a6760547-a35d-4676-9ef9-ca5b94a76fb2-kube-api-access-njcc7\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.040540 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6760547-a35d-4676-9ef9-ca5b94a76fb2-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.317727 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-76bjl" event={"ID":"a6760547-a35d-4676-9ef9-ca5b94a76fb2","Type":"ContainerDied","Data":"92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530"} Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.317776 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92c32ce01df64825c056658c3c1538fbab2b13ac2813d9615c8ed3e49b12e530" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.317798 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-76bjl" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.385595 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-4dhfn"] Dec 12 08:50:31 crc kubenswrapper[4867]: E1212 08:50:31.386186 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6760547-a35d-4676-9ef9-ca5b94a76fb2" containerName="run-os-openstack-openstack-cell1" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.386207 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6760547-a35d-4676-9ef9-ca5b94a76fb2" containerName="run-os-openstack-openstack-cell1" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.386495 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6760547-a35d-4676-9ef9-ca5b94a76fb2" containerName="run-os-openstack-openstack-cell1" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.387293 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.389274 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.389493 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.390193 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.390490 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.395114 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-4dhfn"] Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.549326 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.549663 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9v4c\" (UniqueName: \"kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.549785 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.651465 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9v4c\" (UniqueName: \"kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.651721 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.651789 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.655571 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.656079 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.675208 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9v4c\" (UniqueName: \"kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c\") pod \"reboot-os-openstack-openstack-cell1-4dhfn\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:31 crc kubenswrapper[4867]: I1212 08:50:31.709495 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:32 crc kubenswrapper[4867]: I1212 08:50:32.262708 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-4dhfn"] Dec 12 08:50:32 crc kubenswrapper[4867]: I1212 08:50:32.329914 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" event={"ID":"27e07254-5f63-4c24-996b-178147917e35","Type":"ContainerStarted","Data":"cd9f404b379c0ecc6c9903a4157ad50525b997cefbee7395610be4198a023542"} Dec 12 08:50:33 crc kubenswrapper[4867]: I1212 08:50:33.262927 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:50:34 crc kubenswrapper[4867]: I1212 08:50:34.347942 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" event={"ID":"27e07254-5f63-4c24-996b-178147917e35","Type":"ContainerStarted","Data":"183a2c3abc5305897b538be06938b0b729d72b4d5e69cf308f9592bc278b6397"} Dec 12 08:50:34 crc kubenswrapper[4867]: I1212 08:50:34.363643 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" podStartSLOduration=2.371416082 podStartE2EDuration="3.363628274s" podCreationTimestamp="2025-12-12 08:50:31 +0000 UTC" firstStartedPulling="2025-12-12 08:50:32.26848142 +0000 UTC m=+7319.839862689" lastFinishedPulling="2025-12-12 08:50:33.260693612 +0000 UTC m=+7320.832074881" observedRunningTime="2025-12-12 08:50:34.361842193 +0000 UTC m=+7321.933223462" watchObservedRunningTime="2025-12-12 08:50:34.363628274 +0000 UTC m=+7321.935009543" Dec 12 08:50:48 crc kubenswrapper[4867]: E1212 08:50:48.813206 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27e07254_5f63_4c24_996b_178147917e35.slice/crio-183a2c3abc5305897b538be06938b0b729d72b4d5e69cf308f9592bc278b6397.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27e07254_5f63_4c24_996b_178147917e35.slice/crio-conmon-183a2c3abc5305897b538be06938b0b729d72b4d5e69cf308f9592bc278b6397.scope\": RecentStats: unable to find data in memory cache]" Dec 12 08:50:49 crc kubenswrapper[4867]: I1212 08:50:49.481215 4867 generic.go:334] "Generic (PLEG): container finished" podID="27e07254-5f63-4c24-996b-178147917e35" containerID="183a2c3abc5305897b538be06938b0b729d72b4d5e69cf308f9592bc278b6397" exitCode=0 Dec 12 08:50:49 crc kubenswrapper[4867]: I1212 08:50:49.481362 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" event={"ID":"27e07254-5f63-4c24-996b-178147917e35","Type":"ContainerDied","Data":"183a2c3abc5305897b538be06938b0b729d72b4d5e69cf308f9592bc278b6397"} Dec 12 08:50:50 crc kubenswrapper[4867]: I1212 08:50:50.923685 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.066044 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory\") pod \"27e07254-5f63-4c24-996b-178147917e35\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.066099 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9v4c\" (UniqueName: \"kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c\") pod \"27e07254-5f63-4c24-996b-178147917e35\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.066141 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key\") pod \"27e07254-5f63-4c24-996b-178147917e35\" (UID: \"27e07254-5f63-4c24-996b-178147917e35\") " Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.071175 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c" (OuterVolumeSpecName: "kube-api-access-x9v4c") pod "27e07254-5f63-4c24-996b-178147917e35" (UID: "27e07254-5f63-4c24-996b-178147917e35"). InnerVolumeSpecName "kube-api-access-x9v4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.118150 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27e07254-5f63-4c24-996b-178147917e35" (UID: "27e07254-5f63-4c24-996b-178147917e35"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.119411 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory" (OuterVolumeSpecName: "inventory") pod "27e07254-5f63-4c24-996b-178147917e35" (UID: "27e07254-5f63-4c24-996b-178147917e35"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.169611 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.169912 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9v4c\" (UniqueName: \"kubernetes.io/projected/27e07254-5f63-4c24-996b-178147917e35-kube-api-access-x9v4c\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.170010 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27e07254-5f63-4c24-996b-178147917e35-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.501719 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" event={"ID":"27e07254-5f63-4c24-996b-178147917e35","Type":"ContainerDied","Data":"cd9f404b379c0ecc6c9903a4157ad50525b997cefbee7395610be4198a023542"} Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.501754 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd9f404b379c0ecc6c9903a4157ad50525b997cefbee7395610be4198a023542" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.501796 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-4dhfn" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.582996 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-jxrx8"] Dec 12 08:50:51 crc kubenswrapper[4867]: E1212 08:50:51.583453 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e07254-5f63-4c24-996b-178147917e35" containerName="reboot-os-openstack-openstack-cell1" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.583470 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e07254-5f63-4c24-996b-178147917e35" containerName="reboot-os-openstack-openstack-cell1" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.583700 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e07254-5f63-4c24-996b-178147917e35" containerName="reboot-os-openstack-openstack-cell1" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.584790 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.587764 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.587868 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.587964 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.588300 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.588472 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.588631 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.590352 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.603541 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.618263 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-jxrx8"] Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679438 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679493 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679645 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679697 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679733 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679781 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.679888 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.680148 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwl7b\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.680413 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.680682 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.680866 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.681020 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.681127 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.681256 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.681361 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783260 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwl7b\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783338 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783384 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783419 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783458 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783475 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783495 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783512 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783544 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783566 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783588 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783606 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783624 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783645 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.783667 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.788776 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.788817 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.789428 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.790000 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.791400 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.792558 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.792759 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.792970 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.793069 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.793537 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.797882 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.799188 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.799888 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.802096 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwl7b\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.804367 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-jxrx8\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:51 crc kubenswrapper[4867]: I1212 08:50:51.904983 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:50:52 crc kubenswrapper[4867]: I1212 08:50:52.410567 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-jxrx8"] Dec 12 08:50:52 crc kubenswrapper[4867]: I1212 08:50:52.529447 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" event={"ID":"3ec8f159-fc91-4347-be68-266e1f8b18e1","Type":"ContainerStarted","Data":"4cdb8716687cf62dd01f5de744ab9d1db034c4ec3fa89c037f01e8feb4a075c6"} Dec 12 08:50:53 crc kubenswrapper[4867]: I1212 08:50:53.538333 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" event={"ID":"3ec8f159-fc91-4347-be68-266e1f8b18e1","Type":"ContainerStarted","Data":"e4ef9ad4dfe457ad2f91f4212e7e9e14c9948a5339cfd47d17f1ed9674e9652f"} Dec 12 08:50:53 crc kubenswrapper[4867]: I1212 08:50:53.569065 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" podStartSLOduration=1.901128537 podStartE2EDuration="2.569046055s" podCreationTimestamp="2025-12-12 08:50:51 +0000 UTC" firstStartedPulling="2025-12-12 08:50:52.419832159 +0000 UTC m=+7339.991213428" lastFinishedPulling="2025-12-12 08:50:53.087749667 +0000 UTC m=+7340.659130946" observedRunningTime="2025-12-12 08:50:53.557341463 +0000 UTC m=+7341.128722752" watchObservedRunningTime="2025-12-12 08:50:53.569046055 +0000 UTC m=+7341.140427324" Dec 12 08:51:27 crc kubenswrapper[4867]: I1212 08:51:27.877712 4867 generic.go:334] "Generic (PLEG): container finished" podID="3ec8f159-fc91-4347-be68-266e1f8b18e1" containerID="e4ef9ad4dfe457ad2f91f4212e7e9e14c9948a5339cfd47d17f1ed9674e9652f" exitCode=0 Dec 12 08:51:27 crc kubenswrapper[4867]: I1212 08:51:27.877796 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" event={"ID":"3ec8f159-fc91-4347-be68-266e1f8b18e1","Type":"ContainerDied","Data":"e4ef9ad4dfe457ad2f91f4212e7e9e14c9948a5339cfd47d17f1ed9674e9652f"} Dec 12 08:51:28 crc kubenswrapper[4867]: I1212 08:51:28.991386 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:51:28 crc kubenswrapper[4867]: I1212 08:51:28.991810 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.333326 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516278 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516336 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516380 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516402 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwl7b\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516429 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516463 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516529 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516618 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516636 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516663 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516703 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516723 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516759 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.516831 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle\") pod \"3ec8f159-fc91-4347-be68-266e1f8b18e1\" (UID: \"3ec8f159-fc91-4347-be68-266e1f8b18e1\") " Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.523370 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.523675 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.523941 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524213 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524253 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524394 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b" (OuterVolumeSpecName: "kube-api-access-vwl7b") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "kube-api-access-vwl7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524474 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524847 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.524887 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.525016 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.525913 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.528297 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.533568 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.552322 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory" (OuterVolumeSpecName: "inventory") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.559442 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ec8f159-fc91-4347-be68-266e1f8b18e1" (UID: "3ec8f159-fc91-4347-be68-266e1f8b18e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618549 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618579 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwl7b\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-kube-api-access-vwl7b\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618590 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618601 4867 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618610 4867 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618620 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618632 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618642 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618652 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618662 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618671 4867 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618679 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618687 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618695 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3ec8f159-fc91-4347-be68-266e1f8b18e1-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.618703 4867 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ec8f159-fc91-4347-be68-266e1f8b18e1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.911296 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" event={"ID":"3ec8f159-fc91-4347-be68-266e1f8b18e1","Type":"ContainerDied","Data":"4cdb8716687cf62dd01f5de744ab9d1db034c4ec3fa89c037f01e8feb4a075c6"} Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.911515 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-jxrx8" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.911601 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cdb8716687cf62dd01f5de744ab9d1db034c4ec3fa89c037f01e8feb4a075c6" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.990119 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rtsj8"] Dec 12 08:51:29 crc kubenswrapper[4867]: E1212 08:51:29.990603 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec8f159-fc91-4347-be68-266e1f8b18e1" containerName="install-certs-openstack-openstack-cell1" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.990623 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec8f159-fc91-4347-be68-266e1f8b18e1" containerName="install-certs-openstack-openstack-cell1" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.990870 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec8f159-fc91-4347-be68-266e1f8b18e1" containerName="install-certs-openstack-openstack-cell1" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.991747 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.994421 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.994549 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.994600 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.994605 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:51:29 crc kubenswrapper[4867]: I1212 08:51:29.995171 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.020687 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rtsj8"] Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.130480 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.130809 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.130931 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.131077 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.131192 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cnps\" (UniqueName: \"kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.233176 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.233585 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cnps\" (UniqueName: \"kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.233832 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.234059 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.234211 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.234801 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.238761 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.246911 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.254064 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.257747 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cnps\" (UniqueName: \"kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps\") pod \"ovn-openstack-openstack-cell1-rtsj8\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.314552 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.856089 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rtsj8"] Dec 12 08:51:30 crc kubenswrapper[4867]: I1212 08:51:30.926471 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" event={"ID":"f4807fa4-5151-46f0-88ea-d9878b926bae","Type":"ContainerStarted","Data":"f09ae1ed1a48a2f1c3e09cd50f4d65a4a11aa9800a198fd4765a2fca2fa92c6b"} Dec 12 08:51:31 crc kubenswrapper[4867]: I1212 08:51:31.936607 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" event={"ID":"f4807fa4-5151-46f0-88ea-d9878b926bae","Type":"ContainerStarted","Data":"e288553e87b6a0d87c8b768a3b9b4fe27ae3fa6d3541f812bcffdedd565d1b32"} Dec 12 08:51:31 crc kubenswrapper[4867]: I1212 08:51:31.960115 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" podStartSLOduration=2.295124845 podStartE2EDuration="2.960095945s" podCreationTimestamp="2025-12-12 08:51:29 +0000 UTC" firstStartedPulling="2025-12-12 08:51:30.861508684 +0000 UTC m=+7378.432889953" lastFinishedPulling="2025-12-12 08:51:31.526479784 +0000 UTC m=+7379.097861053" observedRunningTime="2025-12-12 08:51:31.951158448 +0000 UTC m=+7379.522539727" watchObservedRunningTime="2025-12-12 08:51:31.960095945 +0000 UTC m=+7379.531477214" Dec 12 08:51:58 crc kubenswrapper[4867]: I1212 08:51:58.988680 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:51:58 crc kubenswrapper[4867]: I1212 08:51:58.989119 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:52:28 crc kubenswrapper[4867]: I1212 08:52:28.989176 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:52:28 crc kubenswrapper[4867]: I1212 08:52:28.989761 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 08:52:28 crc kubenswrapper[4867]: I1212 08:52:28.989809 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 08:52:28 crc kubenswrapper[4867]: I1212 08:52:28.990654 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 08:52:28 crc kubenswrapper[4867]: I1212 08:52:28.990741 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" gracePeriod=600 Dec 12 08:52:29 crc kubenswrapper[4867]: E1212 08:52:29.114926 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:52:29 crc kubenswrapper[4867]: I1212 08:52:29.512307 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" exitCode=0 Dec 12 08:52:29 crc kubenswrapper[4867]: I1212 08:52:29.512397 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8"} Dec 12 08:52:29 crc kubenswrapper[4867]: I1212 08:52:29.512939 4867 scope.go:117] "RemoveContainer" containerID="80c8d45d13f80784aed9e7c7db449836d75fe5c86511e6756ff1c99a5f63e9f2" Dec 12 08:52:29 crc kubenswrapper[4867]: I1212 08:52:29.513875 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:52:29 crc kubenswrapper[4867]: E1212 08:52:29.514287 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:52:35 crc kubenswrapper[4867]: I1212 08:52:35.573111 4867 generic.go:334] "Generic (PLEG): container finished" podID="f4807fa4-5151-46f0-88ea-d9878b926bae" containerID="e288553e87b6a0d87c8b768a3b9b4fe27ae3fa6d3541f812bcffdedd565d1b32" exitCode=0 Dec 12 08:52:35 crc kubenswrapper[4867]: I1212 08:52:35.573196 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" event={"ID":"f4807fa4-5151-46f0-88ea-d9878b926bae","Type":"ContainerDied","Data":"e288553e87b6a0d87c8b768a3b9b4fe27ae3fa6d3541f812bcffdedd565d1b32"} Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.038190 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.091825 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory\") pod \"f4807fa4-5151-46f0-88ea-d9878b926bae\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.092003 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key\") pod \"f4807fa4-5151-46f0-88ea-d9878b926bae\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.092065 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0\") pod \"f4807fa4-5151-46f0-88ea-d9878b926bae\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.092168 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cnps\" (UniqueName: \"kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps\") pod \"f4807fa4-5151-46f0-88ea-d9878b926bae\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.092253 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle\") pod \"f4807fa4-5151-46f0-88ea-d9878b926bae\" (UID: \"f4807fa4-5151-46f0-88ea-d9878b926bae\") " Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.097952 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps" (OuterVolumeSpecName: "kube-api-access-8cnps") pod "f4807fa4-5151-46f0-88ea-d9878b926bae" (UID: "f4807fa4-5151-46f0-88ea-d9878b926bae"). InnerVolumeSpecName "kube-api-access-8cnps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.099543 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f4807fa4-5151-46f0-88ea-d9878b926bae" (UID: "f4807fa4-5151-46f0-88ea-d9878b926bae"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.122998 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory" (OuterVolumeSpecName: "inventory") pod "f4807fa4-5151-46f0-88ea-d9878b926bae" (UID: "f4807fa4-5151-46f0-88ea-d9878b926bae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.129472 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f4807fa4-5151-46f0-88ea-d9878b926bae" (UID: "f4807fa4-5151-46f0-88ea-d9878b926bae"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.133170 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f4807fa4-5151-46f0-88ea-d9878b926bae" (UID: "f4807fa4-5151-46f0-88ea-d9878b926bae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.195252 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.195285 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.195296 4867 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f4807fa4-5151-46f0-88ea-d9878b926bae-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.195307 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cnps\" (UniqueName: \"kubernetes.io/projected/f4807fa4-5151-46f0-88ea-d9878b926bae-kube-api-access-8cnps\") on node \"crc\" DevicePath \"\"" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.195316 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4807fa4-5151-46f0-88ea-d9878b926bae-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.606013 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" event={"ID":"f4807fa4-5151-46f0-88ea-d9878b926bae","Type":"ContainerDied","Data":"f09ae1ed1a48a2f1c3e09cd50f4d65a4a11aa9800a198fd4765a2fca2fa92c6b"} Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.606408 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f09ae1ed1a48a2f1c3e09cd50f4d65a4a11aa9800a198fd4765a2fca2fa92c6b" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.606059 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rtsj8" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.763593 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-h4zwq"] Dec 12 08:52:37 crc kubenswrapper[4867]: E1212 08:52:37.764099 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4807fa4-5151-46f0-88ea-d9878b926bae" containerName="ovn-openstack-openstack-cell1" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.764123 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4807fa4-5151-46f0-88ea-d9878b926bae" containerName="ovn-openstack-openstack-cell1" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.764415 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4807fa4-5151-46f0-88ea-d9878b926bae" containerName="ovn-openstack-openstack-cell1" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.766622 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.768842 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.769060 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.769772 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.769820 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.774623 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.774695 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.786166 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-h4zwq"] Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.805933 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9grg\" (UniqueName: \"kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.806000 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.806115 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.806275 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.806308 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.806344 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908373 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9grg\" (UniqueName: \"kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908460 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908578 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908656 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908713 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.908745 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.912694 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.913035 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.913941 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.915810 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.916869 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:37 crc kubenswrapper[4867]: I1212 08:52:37.925125 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9grg\" (UniqueName: \"kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg\") pod \"neutron-metadata-openstack-openstack-cell1-h4zwq\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:38 crc kubenswrapper[4867]: I1212 08:52:38.091846 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:52:38 crc kubenswrapper[4867]: I1212 08:52:38.720342 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-h4zwq"] Dec 12 08:52:38 crc kubenswrapper[4867]: W1212 08:52:38.726536 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a40e922_5701_4ecf_b258_c8c129fd6ca1.slice/crio-fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44 WatchSource:0}: Error finding container fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44: Status 404 returned error can't find the container with id fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44 Dec 12 08:52:38 crc kubenswrapper[4867]: I1212 08:52:38.729237 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:52:39 crc kubenswrapper[4867]: I1212 08:52:39.626940 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" event={"ID":"9a40e922-5701-4ecf-b258-c8c129fd6ca1","Type":"ContainerStarted","Data":"fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44"} Dec 12 08:52:40 crc kubenswrapper[4867]: I1212 08:52:40.637209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" event={"ID":"9a40e922-5701-4ecf-b258-c8c129fd6ca1","Type":"ContainerStarted","Data":"009d2b4d81f50aa7c8cefed39c6427f488210bff5b00eb685f20d203fc54bd41"} Dec 12 08:52:40 crc kubenswrapper[4867]: I1212 08:52:40.656794 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" podStartSLOduration=2.725903061 podStartE2EDuration="3.656777674s" podCreationTimestamp="2025-12-12 08:52:37 +0000 UTC" firstStartedPulling="2025-12-12 08:52:38.728877498 +0000 UTC m=+7446.300258767" lastFinishedPulling="2025-12-12 08:52:39.659752111 +0000 UTC m=+7447.231133380" observedRunningTime="2025-12-12 08:52:40.651481474 +0000 UTC m=+7448.222862743" watchObservedRunningTime="2025-12-12 08:52:40.656777674 +0000 UTC m=+7448.228158943" Dec 12 08:52:41 crc kubenswrapper[4867]: I1212 08:52:41.838553 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:52:41 crc kubenswrapper[4867]: E1212 08:52:41.839145 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:52:53 crc kubenswrapper[4867]: I1212 08:52:53.839493 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:52:53 crc kubenswrapper[4867]: E1212 08:52:53.840366 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:53:06 crc kubenswrapper[4867]: I1212 08:53:06.838138 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:53:06 crc kubenswrapper[4867]: E1212 08:53:06.839030 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:53:19 crc kubenswrapper[4867]: I1212 08:53:19.838587 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:53:19 crc kubenswrapper[4867]: E1212 08:53:19.839482 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:53:31 crc kubenswrapper[4867]: I1212 08:53:31.534399 4867 generic.go:334] "Generic (PLEG): container finished" podID="9a40e922-5701-4ecf-b258-c8c129fd6ca1" containerID="009d2b4d81f50aa7c8cefed39c6427f488210bff5b00eb685f20d203fc54bd41" exitCode=0 Dec 12 08:53:31 crc kubenswrapper[4867]: I1212 08:53:31.534447 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" event={"ID":"9a40e922-5701-4ecf-b258-c8c129fd6ca1","Type":"ContainerDied","Data":"009d2b4d81f50aa7c8cefed39c6427f488210bff5b00eb685f20d203fc54bd41"} Dec 12 08:53:32 crc kubenswrapper[4867]: I1212 08:53:32.984450 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105345 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105458 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105502 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105537 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9grg\" (UniqueName: \"kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105563 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.105623 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory\") pod \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\" (UID: \"9a40e922-5701-4ecf-b258-c8c129fd6ca1\") " Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.110585 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.111041 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg" (OuterVolumeSpecName: "kube-api-access-h9grg") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "kube-api-access-h9grg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.137771 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.138345 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.138410 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.139931 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory" (OuterVolumeSpecName: "inventory") pod "9a40e922-5701-4ecf-b258-c8c129fd6ca1" (UID: "9a40e922-5701-4ecf-b258-c8c129fd6ca1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208851 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208893 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208911 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9grg\" (UniqueName: \"kubernetes.io/projected/9a40e922-5701-4ecf-b258-c8c129fd6ca1-kube-api-access-h9grg\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208926 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208939 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.208950 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9a40e922-5701-4ecf-b258-c8c129fd6ca1-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.554241 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.554203 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-h4zwq" event={"ID":"9a40e922-5701-4ecf-b258-c8c129fd6ca1","Type":"ContainerDied","Data":"fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44"} Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.555447 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fafbe7ab50b940d5b3e81d89352cb1f69e1da5991ccfcaa1631630cf723a8e44" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.657708 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-x52kw"] Dec 12 08:53:33 crc kubenswrapper[4867]: E1212 08:53:33.658120 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a40e922-5701-4ecf-b258-c8c129fd6ca1" containerName="neutron-metadata-openstack-openstack-cell1" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.658138 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a40e922-5701-4ecf-b258-c8c129fd6ca1" containerName="neutron-metadata-openstack-openstack-cell1" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.668334 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a40e922-5701-4ecf-b258-c8c129fd6ca1" containerName="neutron-metadata-openstack-openstack-cell1" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.669065 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-x52kw"] Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.669170 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.672515 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.690257 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.690489 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.690630 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.690757 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.821851 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.822676 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5vgk\" (UniqueName: \"kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.822837 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.822997 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.823121 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.838375 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:53:33 crc kubenswrapper[4867]: E1212 08:53:33.838752 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.924825 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5vgk\" (UniqueName: \"kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.924901 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.924958 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.924977 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.925106 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.929265 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.929419 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.930001 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.931260 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.941943 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5vgk\" (UniqueName: \"kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk\") pod \"libvirt-openstack-openstack-cell1-x52kw\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:33 crc kubenswrapper[4867]: I1212 08:53:33.991105 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:53:34 crc kubenswrapper[4867]: I1212 08:53:34.539336 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-x52kw"] Dec 12 08:53:34 crc kubenswrapper[4867]: W1212 08:53:34.548410 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cdd20a7_5b34_4ab7_a93a_8b8f273535e7.slice/crio-f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087 WatchSource:0}: Error finding container f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087: Status 404 returned error can't find the container with id f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087 Dec 12 08:53:34 crc kubenswrapper[4867]: I1212 08:53:34.565946 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" event={"ID":"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7","Type":"ContainerStarted","Data":"f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087"} Dec 12 08:53:35 crc kubenswrapper[4867]: I1212 08:53:35.576073 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" event={"ID":"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7","Type":"ContainerStarted","Data":"ac920e00ecce0e6b8460e793dd56616603224f204d2f9aca75e644fab9e88e4f"} Dec 12 08:53:35 crc kubenswrapper[4867]: I1212 08:53:35.601110 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" podStartSLOduration=2.134861142 podStartE2EDuration="2.601090093s" podCreationTimestamp="2025-12-12 08:53:33 +0000 UTC" firstStartedPulling="2025-12-12 08:53:34.553040154 +0000 UTC m=+7502.124421423" lastFinishedPulling="2025-12-12 08:53:35.019269105 +0000 UTC m=+7502.590650374" observedRunningTime="2025-12-12 08:53:35.59566096 +0000 UTC m=+7503.167042229" watchObservedRunningTime="2025-12-12 08:53:35.601090093 +0000 UTC m=+7503.172471362" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.719404 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.722722 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.733972 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.819110 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.819297 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.819322 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9htj\" (UniqueName: \"kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.920660 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.921055 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.921073 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9htj\" (UniqueName: \"kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.921917 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.922140 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:42 crc kubenswrapper[4867]: I1212 08:53:42.941274 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9htj\" (UniqueName: \"kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj\") pod \"redhat-marketplace-g6lhp\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:43 crc kubenswrapper[4867]: I1212 08:53:43.042560 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:43 crc kubenswrapper[4867]: I1212 08:53:43.624251 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:43 crc kubenswrapper[4867]: I1212 08:53:43.655465 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerStarted","Data":"0492867fc77886f2656683bd2a1c98eab6a232d9fefea232195709e613ef043e"} Dec 12 08:53:44 crc kubenswrapper[4867]: I1212 08:53:44.668773 4867 generic.go:334] "Generic (PLEG): container finished" podID="26a09b39-90fe-494e-8c44-6e41796302af" containerID="90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10" exitCode=0 Dec 12 08:53:44 crc kubenswrapper[4867]: I1212 08:53:44.668834 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerDied","Data":"90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10"} Dec 12 08:53:44 crc kubenswrapper[4867]: I1212 08:53:44.839018 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:53:44 crc kubenswrapper[4867]: E1212 08:53:44.839593 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:53:45 crc kubenswrapper[4867]: I1212 08:53:45.681167 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerStarted","Data":"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248"} Dec 12 08:53:46 crc kubenswrapper[4867]: I1212 08:53:46.691552 4867 generic.go:334] "Generic (PLEG): container finished" podID="26a09b39-90fe-494e-8c44-6e41796302af" containerID="1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248" exitCode=0 Dec 12 08:53:46 crc kubenswrapper[4867]: I1212 08:53:46.692305 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerDied","Data":"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248"} Dec 12 08:53:47 crc kubenswrapper[4867]: I1212 08:53:47.702329 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerStarted","Data":"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb"} Dec 12 08:53:47 crc kubenswrapper[4867]: I1212 08:53:47.724568 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g6lhp" podStartSLOduration=3.29585605 podStartE2EDuration="5.72455229s" podCreationTimestamp="2025-12-12 08:53:42 +0000 UTC" firstStartedPulling="2025-12-12 08:53:44.670987392 +0000 UTC m=+7512.242368651" lastFinishedPulling="2025-12-12 08:53:47.099683622 +0000 UTC m=+7514.671064891" observedRunningTime="2025-12-12 08:53:47.722273374 +0000 UTC m=+7515.293654643" watchObservedRunningTime="2025-12-12 08:53:47.72455229 +0000 UTC m=+7515.295933559" Dec 12 08:53:53 crc kubenswrapper[4867]: I1212 08:53:53.042815 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:53 crc kubenswrapper[4867]: I1212 08:53:53.044324 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:53 crc kubenswrapper[4867]: I1212 08:53:53.097942 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:53 crc kubenswrapper[4867]: I1212 08:53:53.821700 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:53 crc kubenswrapper[4867]: I1212 08:53:53.867047 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:55 crc kubenswrapper[4867]: I1212 08:53:55.790757 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g6lhp" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="registry-server" containerID="cri-o://4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb" gracePeriod=2 Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.302722 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.376959 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities\") pod \"26a09b39-90fe-494e-8c44-6e41796302af\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.377076 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content\") pod \"26a09b39-90fe-494e-8c44-6e41796302af\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.377112 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9htj\" (UniqueName: \"kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj\") pod \"26a09b39-90fe-494e-8c44-6e41796302af\" (UID: \"26a09b39-90fe-494e-8c44-6e41796302af\") " Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.378453 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities" (OuterVolumeSpecName: "utilities") pod "26a09b39-90fe-494e-8c44-6e41796302af" (UID: "26a09b39-90fe-494e-8c44-6e41796302af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.389067 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj" (OuterVolumeSpecName: "kube-api-access-p9htj") pod "26a09b39-90fe-494e-8c44-6e41796302af" (UID: "26a09b39-90fe-494e-8c44-6e41796302af"). InnerVolumeSpecName "kube-api-access-p9htj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.401618 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26a09b39-90fe-494e-8c44-6e41796302af" (UID: "26a09b39-90fe-494e-8c44-6e41796302af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.479637 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.479671 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26a09b39-90fe-494e-8c44-6e41796302af-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.479682 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9htj\" (UniqueName: \"kubernetes.io/projected/26a09b39-90fe-494e-8c44-6e41796302af-kube-api-access-p9htj\") on node \"crc\" DevicePath \"\"" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.800709 4867 generic.go:334] "Generic (PLEG): container finished" podID="26a09b39-90fe-494e-8c44-6e41796302af" containerID="4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb" exitCode=0 Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.800751 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g6lhp" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.800793 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerDied","Data":"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb"} Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.801197 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g6lhp" event={"ID":"26a09b39-90fe-494e-8c44-6e41796302af","Type":"ContainerDied","Data":"0492867fc77886f2656683bd2a1c98eab6a232d9fefea232195709e613ef043e"} Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.801258 4867 scope.go:117] "RemoveContainer" containerID="4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.832809 4867 scope.go:117] "RemoveContainer" containerID="1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.851867 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.864509 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g6lhp"] Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.868507 4867 scope.go:117] "RemoveContainer" containerID="90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.907985 4867 scope.go:117] "RemoveContainer" containerID="4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb" Dec 12 08:53:56 crc kubenswrapper[4867]: E1212 08:53:56.908502 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb\": container with ID starting with 4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb not found: ID does not exist" containerID="4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.908599 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb"} err="failed to get container status \"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb\": rpc error: code = NotFound desc = could not find container \"4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb\": container with ID starting with 4f68b35305e268919be3b5c37e0349bd8f5e7e7cf86853e710ea773a50a88ceb not found: ID does not exist" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.908641 4867 scope.go:117] "RemoveContainer" containerID="1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248" Dec 12 08:53:56 crc kubenswrapper[4867]: E1212 08:53:56.909115 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248\": container with ID starting with 1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248 not found: ID does not exist" containerID="1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.909149 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248"} err="failed to get container status \"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248\": rpc error: code = NotFound desc = could not find container \"1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248\": container with ID starting with 1c3cbeca3d8602624844e87efadd645c5176b7dc90c64de7672222158f857248 not found: ID does not exist" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.909168 4867 scope.go:117] "RemoveContainer" containerID="90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10" Dec 12 08:53:56 crc kubenswrapper[4867]: E1212 08:53:56.909465 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10\": container with ID starting with 90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10 not found: ID does not exist" containerID="90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10" Dec 12 08:53:56 crc kubenswrapper[4867]: I1212 08:53:56.909491 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10"} err="failed to get container status \"90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10\": rpc error: code = NotFound desc = could not find container \"90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10\": container with ID starting with 90caf14556af2d7d12757c1c78e1fbad187cfd5bdc6bae74b2642117ba3e7d10 not found: ID does not exist" Dec 12 08:53:58 crc kubenswrapper[4867]: I1212 08:53:58.847881 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a09b39-90fe-494e-8c44-6e41796302af" path="/var/lib/kubelet/pods/26a09b39-90fe-494e-8c44-6e41796302af/volumes" Dec 12 08:53:59 crc kubenswrapper[4867]: I1212 08:53:59.838400 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:53:59 crc kubenswrapper[4867]: E1212 08:53:59.839092 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:54:11 crc kubenswrapper[4867]: I1212 08:54:11.838941 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:54:11 crc kubenswrapper[4867]: E1212 08:54:11.839888 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:54:23 crc kubenswrapper[4867]: I1212 08:54:23.838582 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:54:23 crc kubenswrapper[4867]: E1212 08:54:23.839796 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:54:35 crc kubenswrapper[4867]: I1212 08:54:35.839178 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:54:35 crc kubenswrapper[4867]: E1212 08:54:35.840087 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:54:47 crc kubenswrapper[4867]: I1212 08:54:47.839453 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:54:47 crc kubenswrapper[4867]: E1212 08:54:47.840476 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:55:02 crc kubenswrapper[4867]: I1212 08:55:02.848150 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:55:02 crc kubenswrapper[4867]: E1212 08:55:02.849352 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:55:13 crc kubenswrapper[4867]: I1212 08:55:13.840650 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:55:13 crc kubenswrapper[4867]: E1212 08:55:13.842293 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:55:24 crc kubenswrapper[4867]: I1212 08:55:24.839058 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:55:24 crc kubenswrapper[4867]: E1212 08:55:24.841196 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:55:36 crc kubenswrapper[4867]: I1212 08:55:36.838393 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:55:36 crc kubenswrapper[4867]: E1212 08:55:36.839108 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:55:49 crc kubenswrapper[4867]: I1212 08:55:49.838445 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:55:49 crc kubenswrapper[4867]: E1212 08:55:49.839446 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:56:03 crc kubenswrapper[4867]: I1212 08:56:03.839131 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:56:03 crc kubenswrapper[4867]: E1212 08:56:03.839861 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:56:15 crc kubenswrapper[4867]: I1212 08:56:15.838836 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:56:15 crc kubenswrapper[4867]: E1212 08:56:15.839891 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:56:29 crc kubenswrapper[4867]: I1212 08:56:29.838135 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:56:29 crc kubenswrapper[4867]: E1212 08:56:29.838948 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.346400 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:33 crc kubenswrapper[4867]: E1212 08:56:33.347129 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="registry-server" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.347141 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="registry-server" Dec 12 08:56:33 crc kubenswrapper[4867]: E1212 08:56:33.347161 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="extract-utilities" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.347169 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="extract-utilities" Dec 12 08:56:33 crc kubenswrapper[4867]: E1212 08:56:33.347182 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="extract-content" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.347189 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="extract-content" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.347372 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a09b39-90fe-494e-8c44-6e41796302af" containerName="registry-server" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.349029 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.356184 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.388113 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.388203 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.388348 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f9q4\" (UniqueName: \"kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.490409 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f9q4\" (UniqueName: \"kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.490485 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.490606 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.491163 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.491576 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.511118 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f9q4\" (UniqueName: \"kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4\") pod \"certified-operators-fj9dj\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:33 crc kubenswrapper[4867]: I1212 08:56:33.673436 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:34 crc kubenswrapper[4867]: I1212 08:56:34.238050 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:34 crc kubenswrapper[4867]: W1212 08:56:34.252935 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podceb0269e_e1b7_4658_a91b_f0a146c7c814.slice/crio-e5cd2ca0d0c5780d260bedaa70c322a90899175f3f7c727e0240716cc642d8eb WatchSource:0}: Error finding container e5cd2ca0d0c5780d260bedaa70c322a90899175f3f7c727e0240716cc642d8eb: Status 404 returned error can't find the container with id e5cd2ca0d0c5780d260bedaa70c322a90899175f3f7c727e0240716cc642d8eb Dec 12 08:56:34 crc kubenswrapper[4867]: I1212 08:56:34.319831 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerStarted","Data":"e5cd2ca0d0c5780d260bedaa70c322a90899175f3f7c727e0240716cc642d8eb"} Dec 12 08:56:35 crc kubenswrapper[4867]: I1212 08:56:35.329268 4867 generic.go:334] "Generic (PLEG): container finished" podID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerID="b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752" exitCode=0 Dec 12 08:56:35 crc kubenswrapper[4867]: I1212 08:56:35.329432 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerDied","Data":"b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752"} Dec 12 08:56:37 crc kubenswrapper[4867]: I1212 08:56:37.349883 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerStarted","Data":"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80"} Dec 12 08:56:38 crc kubenswrapper[4867]: I1212 08:56:38.365838 4867 generic.go:334] "Generic (PLEG): container finished" podID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerID="f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80" exitCode=0 Dec 12 08:56:38 crc kubenswrapper[4867]: I1212 08:56:38.365935 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerDied","Data":"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80"} Dec 12 08:56:39 crc kubenswrapper[4867]: I1212 08:56:39.376488 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerStarted","Data":"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd"} Dec 12 08:56:39 crc kubenswrapper[4867]: I1212 08:56:39.401103 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fj9dj" podStartSLOduration=2.891685669 podStartE2EDuration="6.401083625s" podCreationTimestamp="2025-12-12 08:56:33 +0000 UTC" firstStartedPulling="2025-12-12 08:56:35.33161219 +0000 UTC m=+7682.902993459" lastFinishedPulling="2025-12-12 08:56:38.841010146 +0000 UTC m=+7686.412391415" observedRunningTime="2025-12-12 08:56:39.393390575 +0000 UTC m=+7686.964771854" watchObservedRunningTime="2025-12-12 08:56:39.401083625 +0000 UTC m=+7686.972464894" Dec 12 08:56:43 crc kubenswrapper[4867]: I1212 08:56:43.673622 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:43 crc kubenswrapper[4867]: I1212 08:56:43.674480 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:43 crc kubenswrapper[4867]: I1212 08:56:43.736004 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:44 crc kubenswrapper[4867]: I1212 08:56:44.486905 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:44 crc kubenswrapper[4867]: I1212 08:56:44.532447 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:44 crc kubenswrapper[4867]: I1212 08:56:44.838577 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:56:44 crc kubenswrapper[4867]: E1212 08:56:44.839151 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:56:46 crc kubenswrapper[4867]: I1212 08:56:46.454331 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fj9dj" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="registry-server" containerID="cri-o://f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd" gracePeriod=2 Dec 12 08:56:46 crc kubenswrapper[4867]: I1212 08:56:46.965085 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.075639 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities\") pod \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.075733 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content\") pod \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.075832 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f9q4\" (UniqueName: \"kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4\") pod \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\" (UID: \"ceb0269e-e1b7-4658-a91b-f0a146c7c814\") " Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.077260 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities" (OuterVolumeSpecName: "utilities") pod "ceb0269e-e1b7-4658-a91b-f0a146c7c814" (UID: "ceb0269e-e1b7-4658-a91b-f0a146c7c814"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.082048 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4" (OuterVolumeSpecName: "kube-api-access-7f9q4") pod "ceb0269e-e1b7-4658-a91b-f0a146c7c814" (UID: "ceb0269e-e1b7-4658-a91b-f0a146c7c814"). InnerVolumeSpecName "kube-api-access-7f9q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.145920 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ceb0269e-e1b7-4658-a91b-f0a146c7c814" (UID: "ceb0269e-e1b7-4658-a91b-f0a146c7c814"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.179374 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.179408 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb0269e-e1b7-4658-a91b-f0a146c7c814-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.179421 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f9q4\" (UniqueName: \"kubernetes.io/projected/ceb0269e-e1b7-4658-a91b-f0a146c7c814-kube-api-access-7f9q4\") on node \"crc\" DevicePath \"\"" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.467547 4867 generic.go:334] "Generic (PLEG): container finished" podID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerID="f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd" exitCode=0 Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.467589 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerDied","Data":"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd"} Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.467618 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fj9dj" event={"ID":"ceb0269e-e1b7-4658-a91b-f0a146c7c814","Type":"ContainerDied","Data":"e5cd2ca0d0c5780d260bedaa70c322a90899175f3f7c727e0240716cc642d8eb"} Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.467614 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fj9dj" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.467634 4867 scope.go:117] "RemoveContainer" containerID="f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.497920 4867 scope.go:117] "RemoveContainer" containerID="f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.501089 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.511543 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fj9dj"] Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.519950 4867 scope.go:117] "RemoveContainer" containerID="b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.568293 4867 scope.go:117] "RemoveContainer" containerID="f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd" Dec 12 08:56:47 crc kubenswrapper[4867]: E1212 08:56:47.568722 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd\": container with ID starting with f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd not found: ID does not exist" containerID="f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.568762 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd"} err="failed to get container status \"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd\": rpc error: code = NotFound desc = could not find container \"f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd\": container with ID starting with f7cc9384926102fc296c6d01041a8c89865912c3bece2e85f7747e8cf6eed2dd not found: ID does not exist" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.568787 4867 scope.go:117] "RemoveContainer" containerID="f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80" Dec 12 08:56:47 crc kubenswrapper[4867]: E1212 08:56:47.569125 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80\": container with ID starting with f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80 not found: ID does not exist" containerID="f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.569169 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80"} err="failed to get container status \"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80\": rpc error: code = NotFound desc = could not find container \"f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80\": container with ID starting with f099acdd0bca0a52b43438ae58644badf3e78aed6d174312d8ef1dfe1ba2dc80 not found: ID does not exist" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.569197 4867 scope.go:117] "RemoveContainer" containerID="b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752" Dec 12 08:56:47 crc kubenswrapper[4867]: E1212 08:56:47.569494 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752\": container with ID starting with b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752 not found: ID does not exist" containerID="b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752" Dec 12 08:56:47 crc kubenswrapper[4867]: I1212 08:56:47.569519 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752"} err="failed to get container status \"b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752\": rpc error: code = NotFound desc = could not find container \"b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752\": container with ID starting with b304588dda5af74c861f36c882c93eb8508ac77e0e83bfcb255ef2e311696752 not found: ID does not exist" Dec 12 08:56:48 crc kubenswrapper[4867]: I1212 08:56:48.857180 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" path="/var/lib/kubelet/pods/ceb0269e-e1b7-4658-a91b-f0a146c7c814/volumes" Dec 12 08:56:59 crc kubenswrapper[4867]: I1212 08:56:59.839913 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:56:59 crc kubenswrapper[4867]: E1212 08:56:59.841009 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:57:11 crc kubenswrapper[4867]: I1212 08:57:11.838982 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:57:11 crc kubenswrapper[4867]: E1212 08:57:11.839841 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:57:26 crc kubenswrapper[4867]: I1212 08:57:26.838485 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:57:26 crc kubenswrapper[4867]: E1212 08:57:26.839458 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 08:57:39 crc kubenswrapper[4867]: I1212 08:57:39.839207 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 08:57:40 crc kubenswrapper[4867]: I1212 08:57:40.976015 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa"} Dec 12 08:58:16 crc kubenswrapper[4867]: I1212 08:58:16.294215 4867 generic.go:334] "Generic (PLEG): container finished" podID="3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" containerID="ac920e00ecce0e6b8460e793dd56616603224f204d2f9aca75e644fab9e88e4f" exitCode=0 Dec 12 08:58:16 crc kubenswrapper[4867]: I1212 08:58:16.294264 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" event={"ID":"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7","Type":"ContainerDied","Data":"ac920e00ecce0e6b8460e793dd56616603224f204d2f9aca75e644fab9e88e4f"} Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.747052 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.925759 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5vgk\" (UniqueName: \"kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk\") pod \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.926149 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key\") pod \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.926292 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory\") pod \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.926445 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle\") pod \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.926592 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0\") pod \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\" (UID: \"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7\") " Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.931664 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" (UID: "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.938265 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk" (OuterVolumeSpecName: "kube-api-access-f5vgk") pod "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" (UID: "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7"). InnerVolumeSpecName "kube-api-access-f5vgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.955425 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory" (OuterVolumeSpecName: "inventory") pod "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" (UID: "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.958021 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" (UID: "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:58:17 crc kubenswrapper[4867]: I1212 08:58:17.960477 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" (UID: "3cdd20a7-5b34-4ab7-a93a-8b8f273535e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.029700 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5vgk\" (UniqueName: \"kubernetes.io/projected/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-kube-api-access-f5vgk\") on node \"crc\" DevicePath \"\"" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.029733 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.029742 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.029751 4867 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.029760 4867 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/3cdd20a7-5b34-4ab7-a93a-8b8f273535e7-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.315749 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" event={"ID":"3cdd20a7-5b34-4ab7-a93a-8b8f273535e7","Type":"ContainerDied","Data":"f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087"} Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.315798 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8fc93f35d0f8f6c21934c2618a55a446476c04e3da3ab334ead585e617d9087" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.315875 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-x52kw" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.441296 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qtdct"] Dec 12 08:58:18 crc kubenswrapper[4867]: E1212 08:58:18.442085 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="registry-server" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442102 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="registry-server" Dec 12 08:58:18 crc kubenswrapper[4867]: E1212 08:58:18.442152 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="extract-utilities" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442159 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="extract-utilities" Dec 12 08:58:18 crc kubenswrapper[4867]: E1212 08:58:18.442181 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" containerName="libvirt-openstack-openstack-cell1" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442188 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" containerName="libvirt-openstack-openstack-cell1" Dec 12 08:58:18 crc kubenswrapper[4867]: E1212 08:58:18.442239 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="extract-content" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442246 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="extract-content" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442623 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb0269e-e1b7-4658-a91b-f0a146c7c814" containerName="registry-server" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.442639 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cdd20a7-5b34-4ab7-a93a-8b8f273535e7" containerName="libvirt-openstack-openstack-cell1" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.443621 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.476400 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.476799 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.478165 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.482044 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.495713 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.496917 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.498356 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.535476 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qtdct"] Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.543538 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.543842 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-998m4\" (UniqueName: \"kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.543967 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544154 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544275 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544422 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544539 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544878 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.544982 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.647613 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.648271 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.648998 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.649137 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.649318 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.649415 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.649558 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.650051 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-998m4\" (UniqueName: \"kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.650182 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.652852 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.658037 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.660839 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.661944 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.661895 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.662830 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.665003 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.665443 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.669832 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-998m4\" (UniqueName: \"kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4\") pod \"nova-cell1-openstack-openstack-cell1-qtdct\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:18 crc kubenswrapper[4867]: I1212 08:58:18.807197 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 08:58:19 crc kubenswrapper[4867]: I1212 08:58:19.398845 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qtdct"] Dec 12 08:58:19 crc kubenswrapper[4867]: I1212 08:58:19.407692 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 08:58:20 crc kubenswrapper[4867]: I1212 08:58:20.349404 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" event={"ID":"7840c155-9e2e-4893-b6df-f81ed09aaaeb","Type":"ContainerStarted","Data":"3fa28f17dad16e36d63e6046485eb3258bd934c1df8e0c3aaacfe1d3242b7f63"} Dec 12 08:58:20 crc kubenswrapper[4867]: I1212 08:58:20.349742 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" event={"ID":"7840c155-9e2e-4893-b6df-f81ed09aaaeb","Type":"ContainerStarted","Data":"807aaae4e3bf6ebb64fe817b19a211bf2a991498f31b4ad0117cf6a1024c2963"} Dec 12 08:58:20 crc kubenswrapper[4867]: I1212 08:58:20.382866 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" podStartSLOduration=1.923293828 podStartE2EDuration="2.382844453s" podCreationTimestamp="2025-12-12 08:58:18 +0000 UTC" firstStartedPulling="2025-12-12 08:58:19.407009967 +0000 UTC m=+7786.978391276" lastFinishedPulling="2025-12-12 08:58:19.866560632 +0000 UTC m=+7787.437941901" observedRunningTime="2025-12-12 08:58:20.373708569 +0000 UTC m=+7787.945089838" watchObservedRunningTime="2025-12-12 08:58:20.382844453 +0000 UTC m=+7787.954225722" Dec 12 08:59:58 crc kubenswrapper[4867]: I1212 08:59:58.988783 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 08:59:58 crc kubenswrapper[4867]: I1212 08:59:58.989513 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.173977 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl"] Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.177034 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.182700 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.183455 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.186528 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl"] Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.285881 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.286138 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwvnt\" (UniqueName: \"kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.286226 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.387587 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwvnt\" (UniqueName: \"kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.387681 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.387749 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.388700 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.394671 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.406358 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwvnt\" (UniqueName: \"kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt\") pod \"collect-profiles-29425500-552gl\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.512573 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:00 crc kubenswrapper[4867]: I1212 09:00:00.899497 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl"] Dec 12 09:00:01 crc kubenswrapper[4867]: I1212 09:00:01.303129 4867 generic.go:334] "Generic (PLEG): container finished" podID="dddadd53-c9e0-4524-a315-72157150ca2c" containerID="3225e49e590bccf5d41b9b0cef876c171c4a3e01ce802a7ac3feb3ad903a7a47" exitCode=0 Dec 12 09:00:01 crc kubenswrapper[4867]: I1212 09:00:01.303337 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" event={"ID":"dddadd53-c9e0-4524-a315-72157150ca2c","Type":"ContainerDied","Data":"3225e49e590bccf5d41b9b0cef876c171c4a3e01ce802a7ac3feb3ad903a7a47"} Dec 12 09:00:01 crc kubenswrapper[4867]: I1212 09:00:01.303547 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" event={"ID":"dddadd53-c9e0-4524-a315-72157150ca2c","Type":"ContainerStarted","Data":"be482141ee12fbb4b0f2fbe1a145b1327fc1fe9ab3b516348c3ab5e84ef65d1c"} Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.634192 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.738717 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume\") pod \"dddadd53-c9e0-4524-a315-72157150ca2c\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.738908 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume\") pod \"dddadd53-c9e0-4524-a315-72157150ca2c\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.738938 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwvnt\" (UniqueName: \"kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt\") pod \"dddadd53-c9e0-4524-a315-72157150ca2c\" (UID: \"dddadd53-c9e0-4524-a315-72157150ca2c\") " Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.739562 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume" (OuterVolumeSpecName: "config-volume") pod "dddadd53-c9e0-4524-a315-72157150ca2c" (UID: "dddadd53-c9e0-4524-a315-72157150ca2c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.744323 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dddadd53-c9e0-4524-a315-72157150ca2c" (UID: "dddadd53-c9e0-4524-a315-72157150ca2c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.744535 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt" (OuterVolumeSpecName: "kube-api-access-hwvnt") pod "dddadd53-c9e0-4524-a315-72157150ca2c" (UID: "dddadd53-c9e0-4524-a315-72157150ca2c"). InnerVolumeSpecName "kube-api-access-hwvnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.840958 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dddadd53-c9e0-4524-a315-72157150ca2c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.840997 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwvnt\" (UniqueName: \"kubernetes.io/projected/dddadd53-c9e0-4524-a315-72157150ca2c-kube-api-access-hwvnt\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:02 crc kubenswrapper[4867]: I1212 09:00:02.841006 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dddadd53-c9e0-4524-a315-72157150ca2c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:03 crc kubenswrapper[4867]: E1212 09:00:03.045458 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddddadd53_c9e0_4524_a315_72157150ca2c.slice/crio-be482141ee12fbb4b0f2fbe1a145b1327fc1fe9ab3b516348c3ab5e84ef65d1c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddddadd53_c9e0_4524_a315_72157150ca2c.slice\": RecentStats: unable to find data in memory cache]" Dec 12 09:00:03 crc kubenswrapper[4867]: I1212 09:00:03.323142 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" event={"ID":"dddadd53-c9e0-4524-a315-72157150ca2c","Type":"ContainerDied","Data":"be482141ee12fbb4b0f2fbe1a145b1327fc1fe9ab3b516348c3ab5e84ef65d1c"} Dec 12 09:00:03 crc kubenswrapper[4867]: I1212 09:00:03.323188 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be482141ee12fbb4b0f2fbe1a145b1327fc1fe9ab3b516348c3ab5e84ef65d1c" Dec 12 09:00:03 crc kubenswrapper[4867]: I1212 09:00:03.323260 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl" Dec 12 09:00:03 crc kubenswrapper[4867]: I1212 09:00:03.720577 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v"] Dec 12 09:00:03 crc kubenswrapper[4867]: I1212 09:00:03.731466 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425455-nwl8v"] Dec 12 09:00:04 crc kubenswrapper[4867]: I1212 09:00:04.849476 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28114eb1-01be-4a03-a4db-60f1848c3d05" path="/var/lib/kubelet/pods/28114eb1-01be-4a03-a4db-60f1848c3d05/volumes" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.754833 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:05 crc kubenswrapper[4867]: E1212 09:00:05.755613 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dddadd53-c9e0-4524-a315-72157150ca2c" containerName="collect-profiles" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.755627 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="dddadd53-c9e0-4524-a315-72157150ca2c" containerName="collect-profiles" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.755830 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="dddadd53-c9e0-4524-a315-72157150ca2c" containerName="collect-profiles" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.757355 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.772302 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.902514 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.902672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btxtb\" (UniqueName: \"kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:05 crc kubenswrapper[4867]: I1212 09:00:05.902770 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.005032 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btxtb\" (UniqueName: \"kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.005112 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.005226 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.005793 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.005858 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.028224 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btxtb\" (UniqueName: \"kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb\") pod \"redhat-operators-jkm8m\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.086397 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:06 crc kubenswrapper[4867]: I1212 09:00:06.645979 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:07 crc kubenswrapper[4867]: I1212 09:00:07.355365 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerID="ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351" exitCode=0 Dec 12 09:00:07 crc kubenswrapper[4867]: I1212 09:00:07.355509 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerDied","Data":"ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351"} Dec 12 09:00:07 crc kubenswrapper[4867]: I1212 09:00:07.355639 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerStarted","Data":"a465007a59688c6735f283e332ce00c788576f2ad74037f04ce94051c0d99149"} Dec 12 09:00:09 crc kubenswrapper[4867]: I1212 09:00:09.375640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerStarted","Data":"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4"} Dec 12 09:00:12 crc kubenswrapper[4867]: I1212 09:00:12.405500 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerID="ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4" exitCode=0 Dec 12 09:00:12 crc kubenswrapper[4867]: I1212 09:00:12.405580 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerDied","Data":"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4"} Dec 12 09:00:13 crc kubenswrapper[4867]: I1212 09:00:13.415741 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerStarted","Data":"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce"} Dec 12 09:00:13 crc kubenswrapper[4867]: I1212 09:00:13.435176 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jkm8m" podStartSLOduration=2.893253878 podStartE2EDuration="8.435160246s" podCreationTimestamp="2025-12-12 09:00:05 +0000 UTC" firstStartedPulling="2025-12-12 09:00:07.359173019 +0000 UTC m=+7894.930554288" lastFinishedPulling="2025-12-12 09:00:12.901079387 +0000 UTC m=+7900.472460656" observedRunningTime="2025-12-12 09:00:13.434699035 +0000 UTC m=+7901.006080304" watchObservedRunningTime="2025-12-12 09:00:13.435160246 +0000 UTC m=+7901.006541505" Dec 12 09:00:16 crc kubenswrapper[4867]: I1212 09:00:16.087341 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:16 crc kubenswrapper[4867]: I1212 09:00:16.087899 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:17 crc kubenswrapper[4867]: I1212 09:00:17.130375 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jkm8m" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="registry-server" probeResult="failure" output=< Dec 12 09:00:17 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 09:00:17 crc kubenswrapper[4867]: > Dec 12 09:00:26 crc kubenswrapper[4867]: I1212 09:00:26.137471 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:26 crc kubenswrapper[4867]: I1212 09:00:26.188904 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:26 crc kubenswrapper[4867]: I1212 09:00:26.374600 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:27 crc kubenswrapper[4867]: I1212 09:00:27.529255 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jkm8m" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="registry-server" containerID="cri-o://cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce" gracePeriod=2 Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.043494 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.176585 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content\") pod \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.176700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities\") pod \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.176748 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btxtb\" (UniqueName: \"kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb\") pod \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\" (UID: \"d1d20eb9-6917-4d9d-b351-998bb1a408b6\") " Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.177500 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities" (OuterVolumeSpecName: "utilities") pod "d1d20eb9-6917-4d9d-b351-998bb1a408b6" (UID: "d1d20eb9-6917-4d9d-b351-998bb1a408b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.183013 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb" (OuterVolumeSpecName: "kube-api-access-btxtb") pod "d1d20eb9-6917-4d9d-b351-998bb1a408b6" (UID: "d1d20eb9-6917-4d9d-b351-998bb1a408b6"). InnerVolumeSpecName "kube-api-access-btxtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.279619 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.279953 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btxtb\" (UniqueName: \"kubernetes.io/projected/d1d20eb9-6917-4d9d-b351-998bb1a408b6-kube-api-access-btxtb\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.324557 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d20eb9-6917-4d9d-b351-998bb1a408b6" (UID: "d1d20eb9-6917-4d9d-b351-998bb1a408b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.382294 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d20eb9-6917-4d9d-b351-998bb1a408b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.539914 4867 generic.go:334] "Generic (PLEG): container finished" podID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerID="cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce" exitCode=0 Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.539962 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerDied","Data":"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce"} Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.539990 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jkm8m" event={"ID":"d1d20eb9-6917-4d9d-b351-998bb1a408b6","Type":"ContainerDied","Data":"a465007a59688c6735f283e332ce00c788576f2ad74037f04ce94051c0d99149"} Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.539994 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jkm8m" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.540013 4867 scope.go:117] "RemoveContainer" containerID="cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.561823 4867 scope.go:117] "RemoveContainer" containerID="ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.581306 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.593844 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jkm8m"] Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.609613 4867 scope.go:117] "RemoveContainer" containerID="ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.638085 4867 scope.go:117] "RemoveContainer" containerID="cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce" Dec 12 09:00:28 crc kubenswrapper[4867]: E1212 09:00:28.638523 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce\": container with ID starting with cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce not found: ID does not exist" containerID="cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.638567 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce"} err="failed to get container status \"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce\": rpc error: code = NotFound desc = could not find container \"cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce\": container with ID starting with cc7a62523115faab5cf513b387efa88e7fd1719ec0f38f46c47e40bd69f6bcce not found: ID does not exist" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.638592 4867 scope.go:117] "RemoveContainer" containerID="ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4" Dec 12 09:00:28 crc kubenswrapper[4867]: E1212 09:00:28.639004 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4\": container with ID starting with ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4 not found: ID does not exist" containerID="ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.639037 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4"} err="failed to get container status \"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4\": rpc error: code = NotFound desc = could not find container \"ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4\": container with ID starting with ab07514c598b12ea2a1fcdaf39083bbe3d3384a7aa756706ecb2a63231eb00d4 not found: ID does not exist" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.639059 4867 scope.go:117] "RemoveContainer" containerID="ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351" Dec 12 09:00:28 crc kubenswrapper[4867]: E1212 09:00:28.639363 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351\": container with ID starting with ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351 not found: ID does not exist" containerID="ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.639397 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351"} err="failed to get container status \"ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351\": rpc error: code = NotFound desc = could not find container \"ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351\": container with ID starting with ed09b66d13948359f9f486b8d404881cbf36378e4849510ad0d0ff09db4fe351 not found: ID does not exist" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.849539 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" path="/var/lib/kubelet/pods/d1d20eb9-6917-4d9d-b351-998bb1a408b6/volumes" Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.989337 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:00:28 crc kubenswrapper[4867]: I1212 09:00:28.989656 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:00:58 crc kubenswrapper[4867]: I1212 09:00:58.989102 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:00:58 crc kubenswrapper[4867]: I1212 09:00:58.989603 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:00:58 crc kubenswrapper[4867]: I1212 09:00:58.989695 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:00:58 crc kubenswrapper[4867]: I1212 09:00:58.990814 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:00:58 crc kubenswrapper[4867]: I1212 09:00:58.990884 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa" gracePeriod=600 Dec 12 09:00:59 crc kubenswrapper[4867]: I1212 09:00:59.849695 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa" exitCode=0 Dec 12 09:00:59 crc kubenswrapper[4867]: I1212 09:00:59.849766 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa"} Dec 12 09:00:59 crc kubenswrapper[4867]: I1212 09:00:59.850323 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027"} Dec 12 09:00:59 crc kubenswrapper[4867]: I1212 09:00:59.850351 4867 scope.go:117] "RemoveContainer" containerID="37575420c16c54aa0522f9b6bffdd2839312e27c65fd5d6d85687393d781f6e8" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.152311 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29425501-xj6cb"] Dec 12 09:01:00 crc kubenswrapper[4867]: E1212 09:01:00.152711 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="registry-server" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.152724 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="registry-server" Dec 12 09:01:00 crc kubenswrapper[4867]: E1212 09:01:00.152739 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="extract-utilities" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.152746 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="extract-utilities" Dec 12 09:01:00 crc kubenswrapper[4867]: E1212 09:01:00.152780 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="extract-content" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.152787 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="extract-content" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.152988 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d20eb9-6917-4d9d-b351-998bb1a408b6" containerName="registry-server" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.153655 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.175107 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29425501-xj6cb"] Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.326929 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.327112 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.327173 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpwz\" (UniqueName: \"kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.327216 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.429166 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.429253 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpwz\" (UniqueName: \"kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.429293 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.429385 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.436974 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.439735 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.440508 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.449146 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpwz\" (UniqueName: \"kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz\") pod \"keystone-cron-29425501-xj6cb\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.494161 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:00 crc kubenswrapper[4867]: I1212 09:01:00.955934 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29425501-xj6cb"] Dec 12 09:01:01 crc kubenswrapper[4867]: I1212 09:01:01.883444 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29425501-xj6cb" event={"ID":"670929f8-dbbc-402a-94f1-6d8def78eace","Type":"ContainerStarted","Data":"e2513c35c3b00d3e1d848f4188b982981ec662145b5bfc9c3f9b9ed69c69df81"} Dec 12 09:01:01 crc kubenswrapper[4867]: I1212 09:01:01.884014 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29425501-xj6cb" event={"ID":"670929f8-dbbc-402a-94f1-6d8def78eace","Type":"ContainerStarted","Data":"18a4ea6f3389d78831be3606c53471b3a86d2096349cc719c6b16cd7f1f325ab"} Dec 12 09:01:01 crc kubenswrapper[4867]: I1212 09:01:01.914480 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29425501-xj6cb" podStartSLOduration=1.914451557 podStartE2EDuration="1.914451557s" podCreationTimestamp="2025-12-12 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:01:01.901272782 +0000 UTC m=+7949.472654071" watchObservedRunningTime="2025-12-12 09:01:01.914451557 +0000 UTC m=+7949.485832826" Dec 12 09:01:02 crc kubenswrapper[4867]: I1212 09:01:02.806877 4867 scope.go:117] "RemoveContainer" containerID="69d2dc93878f543632eb1060dad7aedfcfe1ec67c2b8cbee6a5779f1dc007983" Dec 12 09:01:04 crc kubenswrapper[4867]: I1212 09:01:04.916406 4867 generic.go:334] "Generic (PLEG): container finished" podID="670929f8-dbbc-402a-94f1-6d8def78eace" containerID="e2513c35c3b00d3e1d848f4188b982981ec662145b5bfc9c3f9b9ed69c69df81" exitCode=0 Dec 12 09:01:04 crc kubenswrapper[4867]: I1212 09:01:04.916546 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29425501-xj6cb" event={"ID":"670929f8-dbbc-402a-94f1-6d8def78eace","Type":"ContainerDied","Data":"e2513c35c3b00d3e1d848f4188b982981ec662145b5bfc9c3f9b9ed69c69df81"} Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.303471 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.453761 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle\") pod \"670929f8-dbbc-402a-94f1-6d8def78eace\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.453888 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys\") pod \"670929f8-dbbc-402a-94f1-6d8def78eace\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.453934 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data\") pod \"670929f8-dbbc-402a-94f1-6d8def78eace\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.454135 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhpwz\" (UniqueName: \"kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz\") pod \"670929f8-dbbc-402a-94f1-6d8def78eace\" (UID: \"670929f8-dbbc-402a-94f1-6d8def78eace\") " Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.470325 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz" (OuterVolumeSpecName: "kube-api-access-vhpwz") pod "670929f8-dbbc-402a-94f1-6d8def78eace" (UID: "670929f8-dbbc-402a-94f1-6d8def78eace"). InnerVolumeSpecName "kube-api-access-vhpwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.472466 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "670929f8-dbbc-402a-94f1-6d8def78eace" (UID: "670929f8-dbbc-402a-94f1-6d8def78eace"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.488595 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "670929f8-dbbc-402a-94f1-6d8def78eace" (UID: "670929f8-dbbc-402a-94f1-6d8def78eace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.556915 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data" (OuterVolumeSpecName: "config-data") pod "670929f8-dbbc-402a-94f1-6d8def78eace" (UID: "670929f8-dbbc-402a-94f1-6d8def78eace"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.557025 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhpwz\" (UniqueName: \"kubernetes.io/projected/670929f8-dbbc-402a-94f1-6d8def78eace-kube-api-access-vhpwz\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.557078 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.557090 4867 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.658789 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/670929f8-dbbc-402a-94f1-6d8def78eace-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.940032 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29425501-xj6cb" event={"ID":"670929f8-dbbc-402a-94f1-6d8def78eace","Type":"ContainerDied","Data":"18a4ea6f3389d78831be3606c53471b3a86d2096349cc719c6b16cd7f1f325ab"} Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.940069 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18a4ea6f3389d78831be3606c53471b3a86d2096349cc719c6b16cd7f1f325ab" Dec 12 09:01:06 crc kubenswrapper[4867]: I1212 09:01:06.940126 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29425501-xj6cb" Dec 12 09:01:21 crc kubenswrapper[4867]: I1212 09:01:21.078740 4867 generic.go:334] "Generic (PLEG): container finished" podID="7840c155-9e2e-4893-b6df-f81ed09aaaeb" containerID="3fa28f17dad16e36d63e6046485eb3258bd934c1df8e0c3aaacfe1d3242b7f63" exitCode=0 Dec 12 09:01:21 crc kubenswrapper[4867]: I1212 09:01:21.078829 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" event={"ID":"7840c155-9e2e-4893-b6df-f81ed09aaaeb","Type":"ContainerDied","Data":"3fa28f17dad16e36d63e6046485eb3258bd934c1df8e0c3aaacfe1d3242b7f63"} Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.639483 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788041 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788096 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-998m4\" (UniqueName: \"kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788133 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788188 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788300 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788322 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788408 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.788994 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.789495 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0\") pod \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\" (UID: \"7840c155-9e2e-4893-b6df-f81ed09aaaeb\") " Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.794404 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4" (OuterVolumeSpecName: "kube-api-access-998m4") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "kube-api-access-998m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.794569 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.815700 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.819448 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory" (OuterVolumeSpecName: "inventory") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.819711 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.828738 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.830003 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.846290 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.853632 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7840c155-9e2e-4893-b6df-f81ed09aaaeb" (UID: "7840c155-9e2e-4893-b6df-f81ed09aaaeb"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.896674 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897108 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-998m4\" (UniqueName: \"kubernetes.io/projected/7840c155-9e2e-4893-b6df-f81ed09aaaeb-kube-api-access-998m4\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897118 4867 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897127 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897137 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897145 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897153 4867 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897162 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:22 crc kubenswrapper[4867]: I1212 09:01:22.897177 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7840c155-9e2e-4893-b6df-f81ed09aaaeb-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.100196 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" event={"ID":"7840c155-9e2e-4893-b6df-f81ed09aaaeb","Type":"ContainerDied","Data":"807aaae4e3bf6ebb64fe817b19a211bf2a991498f31b4ad0117cf6a1024c2963"} Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.100248 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807aaae4e3bf6ebb64fe817b19a211bf2a991498f31b4ad0117cf6a1024c2963" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.100272 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qtdct" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.221277 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-nnz9m"] Dec 12 09:01:23 crc kubenswrapper[4867]: E1212 09:01:23.221888 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670929f8-dbbc-402a-94f1-6d8def78eace" containerName="keystone-cron" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.221913 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="670929f8-dbbc-402a-94f1-6d8def78eace" containerName="keystone-cron" Dec 12 09:01:23 crc kubenswrapper[4867]: E1212 09:01:23.221944 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7840c155-9e2e-4893-b6df-f81ed09aaaeb" containerName="nova-cell1-openstack-openstack-cell1" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.221954 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7840c155-9e2e-4893-b6df-f81ed09aaaeb" containerName="nova-cell1-openstack-openstack-cell1" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.222289 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7840c155-9e2e-4893-b6df-f81ed09aaaeb" containerName="nova-cell1-openstack-openstack-cell1" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.222318 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="670929f8-dbbc-402a-94f1-6d8def78eace" containerName="keystone-cron" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.223178 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.226659 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.226687 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.226717 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.226659 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.227777 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.233703 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-nnz9m"] Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.303765 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.303845 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.303896 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlmsj\" (UniqueName: \"kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.303989 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.304022 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.304055 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.304124 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406390 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406486 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406545 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlmsj\" (UniqueName: \"kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406641 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406664 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406690 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.406754 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.410623 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.410623 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.410972 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.411775 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.412014 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.417473 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.425110 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlmsj\" (UniqueName: \"kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj\") pod \"telemetry-openstack-openstack-cell1-nnz9m\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:23 crc kubenswrapper[4867]: I1212 09:01:23.541792 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:01:24 crc kubenswrapper[4867]: I1212 09:01:24.082886 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-nnz9m"] Dec 12 09:01:24 crc kubenswrapper[4867]: I1212 09:01:24.110565 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" event={"ID":"9cb08d58-7abd-4896-975d-aa575c651851","Type":"ContainerStarted","Data":"f393af464c36b9b315a7b5b67391b0b48291f9195aae6007e49b7eb36dc5a225"} Dec 12 09:01:26 crc kubenswrapper[4867]: I1212 09:01:26.129648 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" event={"ID":"9cb08d58-7abd-4896-975d-aa575c651851","Type":"ContainerStarted","Data":"cb274a7316486dff877b1074e19d13580ef0a0704ea3185131518940a96e7b1a"} Dec 12 09:01:26 crc kubenswrapper[4867]: I1212 09:01:26.154732 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" podStartSLOduration=1.87548209 podStartE2EDuration="3.154715217s" podCreationTimestamp="2025-12-12 09:01:23 +0000 UTC" firstStartedPulling="2025-12-12 09:01:24.08671208 +0000 UTC m=+7971.658093359" lastFinishedPulling="2025-12-12 09:01:25.365945217 +0000 UTC m=+7972.937326486" observedRunningTime="2025-12-12 09:01:26.150755379 +0000 UTC m=+7973.722136668" watchObservedRunningTime="2025-12-12 09:01:26.154715217 +0000 UTC m=+7973.726096486" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.525796 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.528568 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.543684 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.603796 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.603879 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn5dz\" (UniqueName: \"kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.604169 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.705864 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.705977 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn5dz\" (UniqueName: \"kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.706084 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.706869 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.707168 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.730405 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn5dz\" (UniqueName: \"kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz\") pod \"community-operators-xnbjt\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:02 crc kubenswrapper[4867]: I1212 09:02:02.856067 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:03 crc kubenswrapper[4867]: I1212 09:02:03.467058 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:03 crc kubenswrapper[4867]: I1212 09:02:03.915728 4867 generic.go:334] "Generic (PLEG): container finished" podID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerID="a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1" exitCode=0 Dec 12 09:02:03 crc kubenswrapper[4867]: I1212 09:02:03.915823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerDied","Data":"a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1"} Dec 12 09:02:03 crc kubenswrapper[4867]: I1212 09:02:03.916122 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerStarted","Data":"020b3a4256988ab8b61ef9f7669d81fc59357b255d15a3a848aae5e89fd7e609"} Dec 12 09:02:05 crc kubenswrapper[4867]: I1212 09:02:05.938506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerStarted","Data":"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b"} Dec 12 09:02:06 crc kubenswrapper[4867]: E1212 09:02:06.214139 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2292ccb_bb92_410d_89a3_8053baeda05e.slice/crio-conmon-891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2292ccb_bb92_410d_89a3_8053baeda05e.slice/crio-891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b.scope\": RecentStats: unable to find data in memory cache]" Dec 12 09:02:06 crc kubenswrapper[4867]: I1212 09:02:06.950342 4867 generic.go:334] "Generic (PLEG): container finished" podID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerID="891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b" exitCode=0 Dec 12 09:02:06 crc kubenswrapper[4867]: I1212 09:02:06.950434 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerDied","Data":"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b"} Dec 12 09:02:08 crc kubenswrapper[4867]: I1212 09:02:08.970505 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerStarted","Data":"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9"} Dec 12 09:02:08 crc kubenswrapper[4867]: I1212 09:02:08.989705 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xnbjt" podStartSLOduration=3.068476628 podStartE2EDuration="6.989680251s" podCreationTimestamp="2025-12-12 09:02:02 +0000 UTC" firstStartedPulling="2025-12-12 09:02:03.919691043 +0000 UTC m=+8011.491072322" lastFinishedPulling="2025-12-12 09:02:07.840894676 +0000 UTC m=+8015.412275945" observedRunningTime="2025-12-12 09:02:08.98925701 +0000 UTC m=+8016.560638279" watchObservedRunningTime="2025-12-12 09:02:08.989680251 +0000 UTC m=+8016.561061540" Dec 12 09:02:12 crc kubenswrapper[4867]: I1212 09:02:12.858576 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:12 crc kubenswrapper[4867]: I1212 09:02:12.859035 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:12 crc kubenswrapper[4867]: I1212 09:02:12.908960 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:13 crc kubenswrapper[4867]: I1212 09:02:13.050535 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:13 crc kubenswrapper[4867]: I1212 09:02:13.515434 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.024064 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xnbjt" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="registry-server" containerID="cri-o://9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9" gracePeriod=2 Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.535699 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.683557 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities\") pod \"a2292ccb-bb92-410d-89a3-8053baeda05e\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.683648 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn5dz\" (UniqueName: \"kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz\") pod \"a2292ccb-bb92-410d-89a3-8053baeda05e\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.683717 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content\") pod \"a2292ccb-bb92-410d-89a3-8053baeda05e\" (UID: \"a2292ccb-bb92-410d-89a3-8053baeda05e\") " Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.684727 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities" (OuterVolumeSpecName: "utilities") pod "a2292ccb-bb92-410d-89a3-8053baeda05e" (UID: "a2292ccb-bb92-410d-89a3-8053baeda05e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.689437 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz" (OuterVolumeSpecName: "kube-api-access-xn5dz") pod "a2292ccb-bb92-410d-89a3-8053baeda05e" (UID: "a2292ccb-bb92-410d-89a3-8053baeda05e"). InnerVolumeSpecName "kube-api-access-xn5dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.737725 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2292ccb-bb92-410d-89a3-8053baeda05e" (UID: "a2292ccb-bb92-410d-89a3-8053baeda05e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.785606 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.785637 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn5dz\" (UniqueName: \"kubernetes.io/projected/a2292ccb-bb92-410d-89a3-8053baeda05e-kube-api-access-xn5dz\") on node \"crc\" DevicePath \"\"" Dec 12 09:02:15 crc kubenswrapper[4867]: I1212 09:02:15.785648 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2292ccb-bb92-410d-89a3-8053baeda05e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.036054 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnbjt" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.036054 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerDied","Data":"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9"} Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.036333 4867 scope.go:117] "RemoveContainer" containerID="9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.035943 4867 generic.go:334] "Generic (PLEG): container finished" podID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerID="9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9" exitCode=0 Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.039688 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnbjt" event={"ID":"a2292ccb-bb92-410d-89a3-8053baeda05e","Type":"ContainerDied","Data":"020b3a4256988ab8b61ef9f7669d81fc59357b255d15a3a848aae5e89fd7e609"} Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.063521 4867 scope.go:117] "RemoveContainer" containerID="891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.077105 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.083929 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xnbjt"] Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.097132 4867 scope.go:117] "RemoveContainer" containerID="a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.139289 4867 scope.go:117] "RemoveContainer" containerID="9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9" Dec 12 09:02:16 crc kubenswrapper[4867]: E1212 09:02:16.139740 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9\": container with ID starting with 9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9 not found: ID does not exist" containerID="9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.139791 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9"} err="failed to get container status \"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9\": rpc error: code = NotFound desc = could not find container \"9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9\": container with ID starting with 9cc4e9b269d544354ac3d0e917077b9e3bef0a9e369fd56968ab49451a9ee8c9 not found: ID does not exist" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.139815 4867 scope.go:117] "RemoveContainer" containerID="891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b" Dec 12 09:02:16 crc kubenswrapper[4867]: E1212 09:02:16.140267 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b\": container with ID starting with 891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b not found: ID does not exist" containerID="891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.140299 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b"} err="failed to get container status \"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b\": rpc error: code = NotFound desc = could not find container \"891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b\": container with ID starting with 891e7fc600cbc939ee7d233dcb3b6a929b8b8ebe5f5edde96460fcb6f6069e0b not found: ID does not exist" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.140315 4867 scope.go:117] "RemoveContainer" containerID="a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1" Dec 12 09:02:16 crc kubenswrapper[4867]: E1212 09:02:16.140631 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1\": container with ID starting with a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1 not found: ID does not exist" containerID="a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.140658 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1"} err="failed to get container status \"a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1\": rpc error: code = NotFound desc = could not find container \"a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1\": container with ID starting with a6739534631774d4a8dc17244e789fca118ebe5ccd7015a76efd64f37a5034e1 not found: ID does not exist" Dec 12 09:02:16 crc kubenswrapper[4867]: I1212 09:02:16.853093 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" path="/var/lib/kubelet/pods/a2292ccb-bb92-410d-89a3-8053baeda05e/volumes" Dec 12 09:03:28 crc kubenswrapper[4867]: I1212 09:03:28.989576 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:03:28 crc kubenswrapper[4867]: I1212 09:03:28.990731 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.019094 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:03:54 crc kubenswrapper[4867]: E1212 09:03:54.019931 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="extract-utilities" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.019943 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="extract-utilities" Dec 12 09:03:54 crc kubenswrapper[4867]: E1212 09:03:54.019985 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="registry-server" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.019991 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="registry-server" Dec 12 09:03:54 crc kubenswrapper[4867]: E1212 09:03:54.020008 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="extract-content" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.020013 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="extract-content" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.020213 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2292ccb-bb92-410d-89a3-8053baeda05e" containerName="registry-server" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.027820 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.037000 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.140559 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndvlx\" (UniqueName: \"kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.140724 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.140819 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.243398 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.243522 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.243605 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndvlx\" (UniqueName: \"kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.243901 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.244248 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.267470 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndvlx\" (UniqueName: \"kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx\") pod \"redhat-marketplace-bx25b\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.360685 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.829705 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:03:54 crc kubenswrapper[4867]: I1212 09:03:54.981823 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerStarted","Data":"a44a26cb04c0cf2645b5b3899ed1ed0b9d6a70372586860d52139326358e7c31"} Dec 12 09:03:55 crc kubenswrapper[4867]: I1212 09:03:55.992081 4867 generic.go:334] "Generic (PLEG): container finished" podID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerID="3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49" exitCode=0 Dec 12 09:03:55 crc kubenswrapper[4867]: I1212 09:03:55.992165 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerDied","Data":"3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49"} Dec 12 09:03:55 crc kubenswrapper[4867]: I1212 09:03:55.994186 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:03:57 crc kubenswrapper[4867]: I1212 09:03:57.004895 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerStarted","Data":"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf"} Dec 12 09:03:58 crc kubenswrapper[4867]: I1212 09:03:58.015754 4867 generic.go:334] "Generic (PLEG): container finished" podID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerID="6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf" exitCode=0 Dec 12 09:03:58 crc kubenswrapper[4867]: I1212 09:03:58.015830 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerDied","Data":"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf"} Dec 12 09:03:58 crc kubenswrapper[4867]: I1212 09:03:58.988577 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:03:58 crc kubenswrapper[4867]: I1212 09:03:58.988877 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:03:59 crc kubenswrapper[4867]: I1212 09:03:59.026540 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerStarted","Data":"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb"} Dec 12 09:03:59 crc kubenswrapper[4867]: I1212 09:03:59.050468 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bx25b" podStartSLOduration=3.488183987 podStartE2EDuration="6.050448408s" podCreationTimestamp="2025-12-12 09:03:53 +0000 UTC" firstStartedPulling="2025-12-12 09:03:55.993988053 +0000 UTC m=+8123.565369322" lastFinishedPulling="2025-12-12 09:03:58.556252474 +0000 UTC m=+8126.127633743" observedRunningTime="2025-12-12 09:03:59.047155518 +0000 UTC m=+8126.618536827" watchObservedRunningTime="2025-12-12 09:03:59.050448408 +0000 UTC m=+8126.621829677" Dec 12 09:04:04 crc kubenswrapper[4867]: I1212 09:04:04.360876 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:04 crc kubenswrapper[4867]: I1212 09:04:04.361566 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:04 crc kubenswrapper[4867]: I1212 09:04:04.406239 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:05 crc kubenswrapper[4867]: I1212 09:04:05.129846 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:05 crc kubenswrapper[4867]: I1212 09:04:05.222670 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.103415 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bx25b" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="registry-server" containerID="cri-o://c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb" gracePeriod=2 Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.618864 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.717135 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndvlx\" (UniqueName: \"kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx\") pod \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.717212 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content\") pod \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.717429 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities\") pod \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\" (UID: \"61784ef4-7348-4486-8ffe-aa99c9fe35fd\") " Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.718167 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities" (OuterVolumeSpecName: "utilities") pod "61784ef4-7348-4486-8ffe-aa99c9fe35fd" (UID: "61784ef4-7348-4486-8ffe-aa99c9fe35fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.723276 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx" (OuterVolumeSpecName: "kube-api-access-ndvlx") pod "61784ef4-7348-4486-8ffe-aa99c9fe35fd" (UID: "61784ef4-7348-4486-8ffe-aa99c9fe35fd"). InnerVolumeSpecName "kube-api-access-ndvlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.743572 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61784ef4-7348-4486-8ffe-aa99c9fe35fd" (UID: "61784ef4-7348-4486-8ffe-aa99c9fe35fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.820047 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndvlx\" (UniqueName: \"kubernetes.io/projected/61784ef4-7348-4486-8ffe-aa99c9fe35fd-kube-api-access-ndvlx\") on node \"crc\" DevicePath \"\"" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.820097 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:04:07 crc kubenswrapper[4867]: I1212 09:04:07.820109 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61784ef4-7348-4486-8ffe-aa99c9fe35fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.120940 4867 generic.go:334] "Generic (PLEG): container finished" podID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerID="c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb" exitCode=0 Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.120999 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerDied","Data":"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb"} Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.121028 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx25b" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.121053 4867 scope.go:117] "RemoveContainer" containerID="c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.121037 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx25b" event={"ID":"61784ef4-7348-4486-8ffe-aa99c9fe35fd","Type":"ContainerDied","Data":"a44a26cb04c0cf2645b5b3899ed1ed0b9d6a70372586860d52139326358e7c31"} Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.157590 4867 scope.go:117] "RemoveContainer" containerID="6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.159195 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.169178 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx25b"] Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.198727 4867 scope.go:117] "RemoveContainer" containerID="3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.223385 4867 scope.go:117] "RemoveContainer" containerID="c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb" Dec 12 09:04:08 crc kubenswrapper[4867]: E1212 09:04:08.223830 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb\": container with ID starting with c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb not found: ID does not exist" containerID="c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.223855 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb"} err="failed to get container status \"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb\": rpc error: code = NotFound desc = could not find container \"c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb\": container with ID starting with c83296115c9c70395b67519e02872fab57ebaf3dcde911474fe18ef9a6c322fb not found: ID does not exist" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.223877 4867 scope.go:117] "RemoveContainer" containerID="6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf" Dec 12 09:04:08 crc kubenswrapper[4867]: E1212 09:04:08.224121 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf\": container with ID starting with 6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf not found: ID does not exist" containerID="6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.224141 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf"} err="failed to get container status \"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf\": rpc error: code = NotFound desc = could not find container \"6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf\": container with ID starting with 6ea3e446b74f0a2575c308d6fee166ad1b6a5c45c65251210a6fde276a4cb6cf not found: ID does not exist" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.224153 4867 scope.go:117] "RemoveContainer" containerID="3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49" Dec 12 09:04:08 crc kubenswrapper[4867]: E1212 09:04:08.224413 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49\": container with ID starting with 3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49 not found: ID does not exist" containerID="3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.224428 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49"} err="failed to get container status \"3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49\": rpc error: code = NotFound desc = could not find container \"3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49\": container with ID starting with 3fcabf2d7248b758d3a3b871890681b0f939d95c47014265d9c453f45c0e8e49 not found: ID does not exist" Dec 12 09:04:08 crc kubenswrapper[4867]: I1212 09:04:08.848368 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" path="/var/lib/kubelet/pods/61784ef4-7348-4486-8ffe-aa99c9fe35fd/volumes" Dec 12 09:04:28 crc kubenswrapper[4867]: I1212 09:04:28.988922 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:04:28 crc kubenswrapper[4867]: I1212 09:04:28.989538 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:04:28 crc kubenswrapper[4867]: I1212 09:04:28.989586 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:04:28 crc kubenswrapper[4867]: I1212 09:04:28.990452 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:04:28 crc kubenswrapper[4867]: I1212 09:04:28.990518 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" gracePeriod=600 Dec 12 09:04:29 crc kubenswrapper[4867]: E1212 09:04:29.123559 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:04:29 crc kubenswrapper[4867]: I1212 09:04:29.364342 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" exitCode=0 Dec 12 09:04:29 crc kubenswrapper[4867]: I1212 09:04:29.364606 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027"} Dec 12 09:04:29 crc kubenswrapper[4867]: I1212 09:04:29.364637 4867 scope.go:117] "RemoveContainer" containerID="ecceeb672bd3b535e240f198915b342aa42f20b7c8580381187d5fb89d6fafaa" Dec 12 09:04:29 crc kubenswrapper[4867]: I1212 09:04:29.365301 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:04:29 crc kubenswrapper[4867]: E1212 09:04:29.366021 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:04:40 crc kubenswrapper[4867]: I1212 09:04:40.839942 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:04:40 crc kubenswrapper[4867]: E1212 09:04:40.841542 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:04:53 crc kubenswrapper[4867]: I1212 09:04:53.838040 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:04:53 crc kubenswrapper[4867]: E1212 09:04:53.838957 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:05:06 crc kubenswrapper[4867]: I1212 09:05:06.839302 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:05:06 crc kubenswrapper[4867]: E1212 09:05:06.843693 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:05:17 crc kubenswrapper[4867]: I1212 09:05:17.836698 4867 generic.go:334] "Generic (PLEG): container finished" podID="9cb08d58-7abd-4896-975d-aa575c651851" containerID="cb274a7316486dff877b1074e19d13580ef0a0704ea3185131518940a96e7b1a" exitCode=0 Dec 12 09:05:17 crc kubenswrapper[4867]: I1212 09:05:17.836804 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" event={"ID":"9cb08d58-7abd-4896-975d-aa575c651851","Type":"ContainerDied","Data":"cb274a7316486dff877b1074e19d13580ef0a0704ea3185131518940a96e7b1a"} Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.271432 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.415857 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.415964 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.416023 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlmsj\" (UniqueName: \"kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.416094 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.416188 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.416252 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.416319 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory\") pod \"9cb08d58-7abd-4896-975d-aa575c651851\" (UID: \"9cb08d58-7abd-4896-975d-aa575c651851\") " Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.421977 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.422154 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj" (OuterVolumeSpecName: "kube-api-access-jlmsj") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "kube-api-access-jlmsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.445836 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory" (OuterVolumeSpecName: "inventory") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.446108 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.447576 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.455957 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.458526 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9cb08d58-7abd-4896-975d-aa575c651851" (UID: "9cb08d58-7abd-4896-975d-aa575c651851"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518691 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlmsj\" (UniqueName: \"kubernetes.io/projected/9cb08d58-7abd-4896-975d-aa575c651851-kube-api-access-jlmsj\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518728 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518740 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518751 4867 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518761 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518770 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.518778 4867 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9cb08d58-7abd-4896-975d-aa575c651851-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.857126 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" event={"ID":"9cb08d58-7abd-4896-975d-aa575c651851","Type":"ContainerDied","Data":"f393af464c36b9b315a7b5b67391b0b48291f9195aae6007e49b7eb36dc5a225"} Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.857164 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-nnz9m" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.857174 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f393af464c36b9b315a7b5b67391b0b48291f9195aae6007e49b7eb36dc5a225" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963175 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5r8zv"] Dec 12 09:05:19 crc kubenswrapper[4867]: E1212 09:05:19.963608 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="extract-content" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963628 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="extract-content" Dec 12 09:05:19 crc kubenswrapper[4867]: E1212 09:05:19.963659 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="registry-server" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963665 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="registry-server" Dec 12 09:05:19 crc kubenswrapper[4867]: E1212 09:05:19.963678 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="extract-utilities" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963685 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="extract-utilities" Dec 12 09:05:19 crc kubenswrapper[4867]: E1212 09:05:19.963699 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb08d58-7abd-4896-975d-aa575c651851" containerName="telemetry-openstack-openstack-cell1" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963705 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb08d58-7abd-4896-975d-aa575c651851" containerName="telemetry-openstack-openstack-cell1" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963909 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="61784ef4-7348-4486-8ffe-aa99c9fe35fd" containerName="registry-server" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.963930 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb08d58-7abd-4896-975d-aa575c651851" containerName="telemetry-openstack-openstack-cell1" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.964618 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.972406 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.972613 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.972935 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.972948 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.973776 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 09:05:19 crc kubenswrapper[4867]: I1212 09:05:19.976244 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5r8zv"] Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.130396 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.130463 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.130554 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.130576 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sxn8\" (UniqueName: \"kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.130696 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.233690 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.233785 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.233878 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.233906 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sxn8\" (UniqueName: \"kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.233934 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.238328 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.238418 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.238452 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.240716 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.249931 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sxn8\" (UniqueName: \"kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8\") pod \"neutron-sriov-openstack-openstack-cell1-5r8zv\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.283356 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.823160 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-5r8zv"] Dec 12 09:05:20 crc kubenswrapper[4867]: I1212 09:05:20.866446 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" event={"ID":"8dce4a04-d8b6-4c8b-b180-a360add31127","Type":"ContainerStarted","Data":"2bcc298ff8ffbf4add696dbc5a87cb9ceafcb94f03a9fde506ec3675ad824d52"} Dec 12 09:05:21 crc kubenswrapper[4867]: I1212 09:05:21.839177 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:05:21 crc kubenswrapper[4867]: E1212 09:05:21.840091 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:05:21 crc kubenswrapper[4867]: I1212 09:05:21.880209 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" event={"ID":"8dce4a04-d8b6-4c8b-b180-a360add31127","Type":"ContainerStarted","Data":"2db09dffdec5e564e7ea0cdebdfa606b4f36b6eda28faff5292d24ea2f7446bb"} Dec 12 09:05:21 crc kubenswrapper[4867]: I1212 09:05:21.908117 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" podStartSLOduration=2.349736293 podStartE2EDuration="2.908091803s" podCreationTimestamp="2025-12-12 09:05:19 +0000 UTC" firstStartedPulling="2025-12-12 09:05:20.825861723 +0000 UTC m=+8208.397242992" lastFinishedPulling="2025-12-12 09:05:21.384217233 +0000 UTC m=+8208.955598502" observedRunningTime="2025-12-12 09:05:21.895490909 +0000 UTC m=+8209.466872178" watchObservedRunningTime="2025-12-12 09:05:21.908091803 +0000 UTC m=+8209.479473082" Dec 12 09:05:23 crc kubenswrapper[4867]: I1212 09:05:23.346409 4867 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-kgxfk container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 12 09:05:23 crc kubenswrapper[4867]: I1212 09:05:23.346758 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-kgxfk" podUID="545d1312-b5d7-40fa-85c4-d54c8e09c7a2" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.1.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 12 09:05:33 crc kubenswrapper[4867]: I1212 09:05:33.838572 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:05:33 crc kubenswrapper[4867]: E1212 09:05:33.840644 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:05:46 crc kubenswrapper[4867]: I1212 09:05:46.837756 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:05:46 crc kubenswrapper[4867]: E1212 09:05:46.838631 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:05:59 crc kubenswrapper[4867]: I1212 09:05:59.838331 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:05:59 crc kubenswrapper[4867]: E1212 09:05:59.839551 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:06:13 crc kubenswrapper[4867]: I1212 09:06:13.838784 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:06:13 crc kubenswrapper[4867]: E1212 09:06:13.841092 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:06:25 crc kubenswrapper[4867]: I1212 09:06:25.838650 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:06:25 crc kubenswrapper[4867]: E1212 09:06:25.839276 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:06:40 crc kubenswrapper[4867]: I1212 09:06:40.838281 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:06:40 crc kubenswrapper[4867]: E1212 09:06:40.839087 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:06:52 crc kubenswrapper[4867]: I1212 09:06:52.845533 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:06:52 crc kubenswrapper[4867]: E1212 09:06:52.846641 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:07:06 crc kubenswrapper[4867]: I1212 09:07:06.839980 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:07:06 crc kubenswrapper[4867]: E1212 09:07:06.840844 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:07:21 crc kubenswrapper[4867]: I1212 09:07:21.838777 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:07:21 crc kubenswrapper[4867]: E1212 09:07:21.839526 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:07:34 crc kubenswrapper[4867]: I1212 09:07:34.838645 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:07:34 crc kubenswrapper[4867]: E1212 09:07:34.839608 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:07:46 crc kubenswrapper[4867]: I1212 09:07:46.839095 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:07:46 crc kubenswrapper[4867]: E1212 09:07:46.840076 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:00 crc kubenswrapper[4867]: I1212 09:08:00.838439 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:08:00 crc kubenswrapper[4867]: E1212 09:08:00.839207 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:14 crc kubenswrapper[4867]: I1212 09:08:14.838599 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:08:14 crc kubenswrapper[4867]: E1212 09:08:14.839455 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:25 crc kubenswrapper[4867]: I1212 09:08:25.838140 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:08:25 crc kubenswrapper[4867]: E1212 09:08:25.839043 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:37 crc kubenswrapper[4867]: I1212 09:08:37.839652 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:08:37 crc kubenswrapper[4867]: E1212 09:08:37.841350 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:51 crc kubenswrapper[4867]: I1212 09:08:51.839246 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:08:51 crc kubenswrapper[4867]: E1212 09:08:51.841908 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.045749 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.048145 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.069037 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.209563 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn5np\" (UniqueName: \"kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.209610 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.210116 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.314021 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn5np\" (UniqueName: \"kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.314111 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.314204 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.315207 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.315770 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.338692 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn5np\" (UniqueName: \"kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np\") pod \"certified-operators-tf845\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:53 crc kubenswrapper[4867]: I1212 09:08:53.377571 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:08:54 crc kubenswrapper[4867]: I1212 09:08:54.743421 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:08:54 crc kubenswrapper[4867]: I1212 09:08:54.966681 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerStarted","Data":"575ff0c481344f33970bfe08235f96028056555cf7f66d37f755be38a09bb5f3"} Dec 12 09:08:55 crc kubenswrapper[4867]: I1212 09:08:55.981516 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerID="3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea" exitCode=0 Dec 12 09:08:55 crc kubenswrapper[4867]: I1212 09:08:55.981637 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerDied","Data":"3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea"} Dec 12 09:08:58 crc kubenswrapper[4867]: I1212 09:08:58.001344 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerStarted","Data":"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38"} Dec 12 09:08:59 crc kubenswrapper[4867]: I1212 09:08:59.015489 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerID="3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38" exitCode=0 Dec 12 09:08:59 crc kubenswrapper[4867]: I1212 09:08:59.015552 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerDied","Data":"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38"} Dec 12 09:08:59 crc kubenswrapper[4867]: I1212 09:08:59.018139 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:09:00 crc kubenswrapper[4867]: I1212 09:09:00.026772 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerStarted","Data":"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269"} Dec 12 09:09:00 crc kubenswrapper[4867]: I1212 09:09:00.047022 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tf845" podStartSLOduration=3.572103912 podStartE2EDuration="7.047006895s" podCreationTimestamp="2025-12-12 09:08:53 +0000 UTC" firstStartedPulling="2025-12-12 09:08:55.984340653 +0000 UTC m=+8423.555721952" lastFinishedPulling="2025-12-12 09:08:59.459243666 +0000 UTC m=+8427.030624935" observedRunningTime="2025-12-12 09:09:00.045415187 +0000 UTC m=+8427.616796456" watchObservedRunningTime="2025-12-12 09:09:00.047006895 +0000 UTC m=+8427.618388154" Dec 12 09:09:03 crc kubenswrapper[4867]: I1212 09:09:03.378063 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:03 crc kubenswrapper[4867]: I1212 09:09:03.378669 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:03 crc kubenswrapper[4867]: I1212 09:09:03.424075 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:04 crc kubenswrapper[4867]: I1212 09:09:04.119131 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:04 crc kubenswrapper[4867]: I1212 09:09:04.636040 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:09:04 crc kubenswrapper[4867]: I1212 09:09:04.838347 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:09:04 crc kubenswrapper[4867]: E1212 09:09:04.838779 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.092313 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tf845" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="registry-server" containerID="cri-o://131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269" gracePeriod=2 Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.540524 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.587199 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities\") pod \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.587320 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content\") pod \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.587471 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn5np\" (UniqueName: \"kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np\") pod \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\" (UID: \"c7ec64d4-81c2-487a-bf3a-959f08b15b41\") " Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.588111 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities" (OuterVolumeSpecName: "utilities") pod "c7ec64d4-81c2-487a-bf3a-959f08b15b41" (UID: "c7ec64d4-81c2-487a-bf3a-959f08b15b41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.593815 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np" (OuterVolumeSpecName: "kube-api-access-tn5np") pod "c7ec64d4-81c2-487a-bf3a-959f08b15b41" (UID: "c7ec64d4-81c2-487a-bf3a-959f08b15b41"). InnerVolumeSpecName "kube-api-access-tn5np". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.642062 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7ec64d4-81c2-487a-bf3a-959f08b15b41" (UID: "c7ec64d4-81c2-487a-bf3a-959f08b15b41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.689709 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.689742 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn5np\" (UniqueName: \"kubernetes.io/projected/c7ec64d4-81c2-487a-bf3a-959f08b15b41-kube-api-access-tn5np\") on node \"crc\" DevicePath \"\"" Dec 12 09:09:06 crc kubenswrapper[4867]: I1212 09:09:06.689753 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ec64d4-81c2-487a-bf3a-959f08b15b41-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:09:07 crc kubenswrapper[4867]: E1212 09:09:07.016877 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7ec64d4_81c2_487a_bf3a_959f08b15b41.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7ec64d4_81c2_487a_bf3a_959f08b15b41.slice/crio-575ff0c481344f33970bfe08235f96028056555cf7f66d37f755be38a09bb5f3\": RecentStats: unable to find data in memory cache]" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.104035 4867 generic.go:334] "Generic (PLEG): container finished" podID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerID="131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269" exitCode=0 Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.104100 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerDied","Data":"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269"} Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.105057 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tf845" event={"ID":"c7ec64d4-81c2-487a-bf3a-959f08b15b41","Type":"ContainerDied","Data":"575ff0c481344f33970bfe08235f96028056555cf7f66d37f755be38a09bb5f3"} Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.104108 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tf845" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.105127 4867 scope.go:117] "RemoveContainer" containerID="131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.127918 4867 scope.go:117] "RemoveContainer" containerID="3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.132522 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.143040 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tf845"] Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.147403 4867 scope.go:117] "RemoveContainer" containerID="3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.191472 4867 scope.go:117] "RemoveContainer" containerID="131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269" Dec 12 09:09:07 crc kubenswrapper[4867]: E1212 09:09:07.192124 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269\": container with ID starting with 131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269 not found: ID does not exist" containerID="131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.192155 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269"} err="failed to get container status \"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269\": rpc error: code = NotFound desc = could not find container \"131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269\": container with ID starting with 131339647ba11830ff661b5e16c0a7ee855557e2a5f67540029d8420b5d6b269 not found: ID does not exist" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.192175 4867 scope.go:117] "RemoveContainer" containerID="3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38" Dec 12 09:09:07 crc kubenswrapper[4867]: E1212 09:09:07.192569 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38\": container with ID starting with 3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38 not found: ID does not exist" containerID="3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.192594 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38"} err="failed to get container status \"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38\": rpc error: code = NotFound desc = could not find container \"3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38\": container with ID starting with 3ba352d07e2918945255942c1046a0278081993795080fb4a3eac2c0c2685e38 not found: ID does not exist" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.192610 4867 scope.go:117] "RemoveContainer" containerID="3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea" Dec 12 09:09:07 crc kubenswrapper[4867]: E1212 09:09:07.192849 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea\": container with ID starting with 3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea not found: ID does not exist" containerID="3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea" Dec 12 09:09:07 crc kubenswrapper[4867]: I1212 09:09:07.192876 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea"} err="failed to get container status \"3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea\": rpc error: code = NotFound desc = could not find container \"3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea\": container with ID starting with 3b8cd2bafcd6eac8ecd60d8584740158e2d2d21cb882809adba7a67cc8b3e1ea not found: ID does not exist" Dec 12 09:09:08 crc kubenswrapper[4867]: I1212 09:09:08.848235 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" path="/var/lib/kubelet/pods/c7ec64d4-81c2-487a-bf3a-959f08b15b41/volumes" Dec 12 09:09:15 crc kubenswrapper[4867]: I1212 09:09:15.838337 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:09:15 crc kubenswrapper[4867]: E1212 09:09:15.839089 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:09:27 crc kubenswrapper[4867]: I1212 09:09:27.838532 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:09:27 crc kubenswrapper[4867]: E1212 09:09:27.839491 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:09:41 crc kubenswrapper[4867]: I1212 09:09:41.838320 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:09:42 crc kubenswrapper[4867]: I1212 09:09:42.454550 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445"} Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.805258 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:32 crc kubenswrapper[4867]: E1212 09:10:32.806975 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="extract-content" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.806992 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="extract-content" Dec 12 09:10:32 crc kubenswrapper[4867]: E1212 09:10:32.807009 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="extract-utilities" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.807037 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="extract-utilities" Dec 12 09:10:32 crc kubenswrapper[4867]: E1212 09:10:32.807071 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="registry-server" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.807079 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="registry-server" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.807561 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ec64d4-81c2-487a-bf3a-959f08b15b41" containerName="registry-server" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.813077 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.818371 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.819983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8n6\" (UniqueName: \"kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.820053 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.820075 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.922149 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8n6\" (UniqueName: \"kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.922569 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.922599 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.923363 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.923410 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:32 crc kubenswrapper[4867]: I1212 09:10:32.942045 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8n6\" (UniqueName: \"kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6\") pod \"redhat-operators-t6lh2\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:33 crc kubenswrapper[4867]: I1212 09:10:33.152522 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:33 crc kubenswrapper[4867]: I1212 09:10:33.599365 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:33 crc kubenswrapper[4867]: I1212 09:10:33.848318 4867 generic.go:334] "Generic (PLEG): container finished" podID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerID="eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983" exitCode=0 Dec 12 09:10:33 crc kubenswrapper[4867]: I1212 09:10:33.848421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerDied","Data":"eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983"} Dec 12 09:10:33 crc kubenswrapper[4867]: I1212 09:10:33.848660 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerStarted","Data":"d3ed91ef632a6bacbcc47ad8b5eaa276894c63728a9b9b8467f77fca9f2bcdac"} Dec 12 09:10:35 crc kubenswrapper[4867]: I1212 09:10:35.866498 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerStarted","Data":"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b"} Dec 12 09:10:38 crc kubenswrapper[4867]: I1212 09:10:38.897980 4867 generic.go:334] "Generic (PLEG): container finished" podID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerID="57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b" exitCode=0 Dec 12 09:10:38 crc kubenswrapper[4867]: I1212 09:10:38.898060 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerDied","Data":"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b"} Dec 12 09:10:39 crc kubenswrapper[4867]: I1212 09:10:39.907460 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerStarted","Data":"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e"} Dec 12 09:10:39 crc kubenswrapper[4867]: I1212 09:10:39.931010 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t6lh2" podStartSLOduration=2.157266836 podStartE2EDuration="7.930992079s" podCreationTimestamp="2025-12-12 09:10:32 +0000 UTC" firstStartedPulling="2025-12-12 09:10:33.850315826 +0000 UTC m=+8521.421697095" lastFinishedPulling="2025-12-12 09:10:39.624041069 +0000 UTC m=+8527.195422338" observedRunningTime="2025-12-12 09:10:39.926986122 +0000 UTC m=+8527.498367411" watchObservedRunningTime="2025-12-12 09:10:39.930992079 +0000 UTC m=+8527.502373348" Dec 12 09:10:43 crc kubenswrapper[4867]: I1212 09:10:43.153618 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:43 crc kubenswrapper[4867]: I1212 09:10:43.155294 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:44 crc kubenswrapper[4867]: I1212 09:10:44.204907 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t6lh2" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="registry-server" probeResult="failure" output=< Dec 12 09:10:44 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 09:10:44 crc kubenswrapper[4867]: > Dec 12 09:10:53 crc kubenswrapper[4867]: I1212 09:10:53.206177 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:53 crc kubenswrapper[4867]: I1212 09:10:53.259257 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:53 crc kubenswrapper[4867]: I1212 09:10:53.446108 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.050005 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t6lh2" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="registry-server" containerID="cri-o://0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e" gracePeriod=2 Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.547337 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.631945 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content\") pod \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.632025 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp8n6\" (UniqueName: \"kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6\") pod \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.632325 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities\") pod \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\" (UID: \"c6d1ec15-29b4-4e4f-9b17-19417ff7724b\") " Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.633182 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities" (OuterVolumeSpecName: "utilities") pod "c6d1ec15-29b4-4e4f-9b17-19417ff7724b" (UID: "c6d1ec15-29b4-4e4f-9b17-19417ff7724b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.637587 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6" (OuterVolumeSpecName: "kube-api-access-tp8n6") pod "c6d1ec15-29b4-4e4f-9b17-19417ff7724b" (UID: "c6d1ec15-29b4-4e4f-9b17-19417ff7724b"). InnerVolumeSpecName "kube-api-access-tp8n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.735758 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.735818 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp8n6\" (UniqueName: \"kubernetes.io/projected/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-kube-api-access-tp8n6\") on node \"crc\" DevicePath \"\"" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.746898 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6d1ec15-29b4-4e4f-9b17-19417ff7724b" (UID: "c6d1ec15-29b4-4e4f-9b17-19417ff7724b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:10:55 crc kubenswrapper[4867]: I1212 09:10:55.837564 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6d1ec15-29b4-4e4f-9b17-19417ff7724b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.059774 4867 generic.go:334] "Generic (PLEG): container finished" podID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerID="0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e" exitCode=0 Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.059815 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerDied","Data":"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e"} Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.059847 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6lh2" event={"ID":"c6d1ec15-29b4-4e4f-9b17-19417ff7724b","Type":"ContainerDied","Data":"d3ed91ef632a6bacbcc47ad8b5eaa276894c63728a9b9b8467f77fca9f2bcdac"} Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.059864 4867 scope.go:117] "RemoveContainer" containerID="0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.059934 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6lh2" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.081647 4867 scope.go:117] "RemoveContainer" containerID="57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.102285 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.112378 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t6lh2"] Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.135293 4867 scope.go:117] "RemoveContainer" containerID="eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.166953 4867 scope.go:117] "RemoveContainer" containerID="0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e" Dec 12 09:10:56 crc kubenswrapper[4867]: E1212 09:10:56.167507 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e\": container with ID starting with 0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e not found: ID does not exist" containerID="0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.167538 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e"} err="failed to get container status \"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e\": rpc error: code = NotFound desc = could not find container \"0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e\": container with ID starting with 0bd91537a6acf209915acc59c50a69cbabf4254dd9549eb22539e9a303cfeb2e not found: ID does not exist" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.167559 4867 scope.go:117] "RemoveContainer" containerID="57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b" Dec 12 09:10:56 crc kubenswrapper[4867]: E1212 09:10:56.168023 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b\": container with ID starting with 57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b not found: ID does not exist" containerID="57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.168046 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b"} err="failed to get container status \"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b\": rpc error: code = NotFound desc = could not find container \"57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b\": container with ID starting with 57eb0dda3dcae8d13166257c4d8021afeebf11fee4f5ecf04c857f1f0fcf0a0b not found: ID does not exist" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.168058 4867 scope.go:117] "RemoveContainer" containerID="eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983" Dec 12 09:10:56 crc kubenswrapper[4867]: E1212 09:10:56.169490 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983\": container with ID starting with eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983 not found: ID does not exist" containerID="eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.169547 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983"} err="failed to get container status \"eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983\": rpc error: code = NotFound desc = could not find container \"eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983\": container with ID starting with eaacb526f69810403c630ca546c376dd3fd977bff7c253eed1553cd65384c983 not found: ID does not exist" Dec 12 09:10:56 crc kubenswrapper[4867]: I1212 09:10:56.853579 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" path="/var/lib/kubelet/pods/c6d1ec15-29b4-4e4f-9b17-19417ff7724b/volumes" Dec 12 09:11:12 crc kubenswrapper[4867]: I1212 09:11:12.240790 4867 generic.go:334] "Generic (PLEG): container finished" podID="8dce4a04-d8b6-4c8b-b180-a360add31127" containerID="2db09dffdec5e564e7ea0cdebdfa606b4f36b6eda28faff5292d24ea2f7446bb" exitCode=0 Dec 12 09:11:12 crc kubenswrapper[4867]: I1212 09:11:12.240871 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" event={"ID":"8dce4a04-d8b6-4c8b-b180-a360add31127","Type":"ContainerDied","Data":"2db09dffdec5e564e7ea0cdebdfa606b4f36b6eda28faff5292d24ea2f7446bb"} Dec 12 09:11:13 crc kubenswrapper[4867]: I1212 09:11:13.682885 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.803727 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key\") pod \"8dce4a04-d8b6-4c8b-b180-a360add31127\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.804195 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory\") pod \"8dce4a04-d8b6-4c8b-b180-a360add31127\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.804249 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sxn8\" (UniqueName: \"kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8\") pod \"8dce4a04-d8b6-4c8b-b180-a360add31127\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.804349 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0\") pod \"8dce4a04-d8b6-4c8b-b180-a360add31127\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.804432 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle\") pod \"8dce4a04-d8b6-4c8b-b180-a360add31127\" (UID: \"8dce4a04-d8b6-4c8b-b180-a360add31127\") " Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.810593 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "8dce4a04-d8b6-4c8b-b180-a360add31127" (UID: "8dce4a04-d8b6-4c8b-b180-a360add31127"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.820123 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8" (OuterVolumeSpecName: "kube-api-access-6sxn8") pod "8dce4a04-d8b6-4c8b-b180-a360add31127" (UID: "8dce4a04-d8b6-4c8b-b180-a360add31127"). InnerVolumeSpecName "kube-api-access-6sxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.838309 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "8dce4a04-d8b6-4c8b-b180-a360add31127" (UID: "8dce4a04-d8b6-4c8b-b180-a360add31127"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.838875 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8dce4a04-d8b6-4c8b-b180-a360add31127" (UID: "8dce4a04-d8b6-4c8b-b180-a360add31127"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.858472 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory" (OuterVolumeSpecName: "inventory") pod "8dce4a04-d8b6-4c8b-b180-a360add31127" (UID: "8dce4a04-d8b6-4c8b-b180-a360add31127"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.907892 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.907921 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sxn8\" (UniqueName: \"kubernetes.io/projected/8dce4a04-d8b6-4c8b-b180-a360add31127-kube-api-access-6sxn8\") on node \"crc\" DevicePath \"\"" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.907934 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.907945 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:13.907957 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8dce4a04-d8b6-4c8b-b180-a360add31127-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.263382 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" event={"ID":"8dce4a04-d8b6-4c8b-b180-a360add31127","Type":"ContainerDied","Data":"2bcc298ff8ffbf4add696dbc5a87cb9ceafcb94f03a9fde506ec3675ad824d52"} Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.263414 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bcc298ff8ffbf4add696dbc5a87cb9ceafcb94f03a9fde506ec3675ad824d52" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.263473 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-5r8zv" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.368653 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-87rk6"] Dec 12 09:11:14 crc kubenswrapper[4867]: E1212 09:11:14.369065 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dce4a04-d8b6-4c8b-b180-a360add31127" containerName="neutron-sriov-openstack-openstack-cell1" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369083 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dce4a04-d8b6-4c8b-b180-a360add31127" containerName="neutron-sriov-openstack-openstack-cell1" Dec 12 09:11:14 crc kubenswrapper[4867]: E1212 09:11:14.369099 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="registry-server" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369105 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="registry-server" Dec 12 09:11:14 crc kubenswrapper[4867]: E1212 09:11:14.369128 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="extract-utilities" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369136 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="extract-utilities" Dec 12 09:11:14 crc kubenswrapper[4867]: E1212 09:11:14.369153 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="extract-content" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369158 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="extract-content" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369367 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6d1ec15-29b4-4e4f-9b17-19417ff7724b" containerName="registry-server" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.369387 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dce4a04-d8b6-4c8b-b180-a360add31127" containerName="neutron-sriov-openstack-openstack-cell1" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.370107 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.374100 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.374480 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.374709 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.374719 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.374903 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.382903 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-87rk6"] Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.532825 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.533037 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.533094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.533121 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xs9l\" (UniqueName: \"kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.533355 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.634790 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.635126 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.635271 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.635299 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.635332 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xs9l\" (UniqueName: \"kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.640017 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.643811 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.643922 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.644351 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.652716 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xs9l\" (UniqueName: \"kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l\") pod \"neutron-dhcp-openstack-openstack-cell1-87rk6\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:14 crc kubenswrapper[4867]: I1212 09:11:14.742483 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:11:15 crc kubenswrapper[4867]: I1212 09:11:15.329701 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-87rk6"] Dec 12 09:11:16 crc kubenswrapper[4867]: I1212 09:11:16.298848 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" event={"ID":"61a43cf7-2462-4a34-a6f8-4c95a7571df5","Type":"ContainerStarted","Data":"f75e75c901b98c61390c6d22a64209ddab0ea6daf716f0d912708878478120ba"} Dec 12 09:11:16 crc kubenswrapper[4867]: I1212 09:11:16.299267 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" event={"ID":"61a43cf7-2462-4a34-a6f8-4c95a7571df5","Type":"ContainerStarted","Data":"ea6e37de6b81fc1f02dda50fac57be91fee50882219d86051804c13db584121b"} Dec 12 09:11:16 crc kubenswrapper[4867]: I1212 09:11:16.322550 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" podStartSLOduration=1.71666418 podStartE2EDuration="2.322532046s" podCreationTimestamp="2025-12-12 09:11:14 +0000 UTC" firstStartedPulling="2025-12-12 09:11:15.337908019 +0000 UTC m=+8562.909289288" lastFinishedPulling="2025-12-12 09:11:15.943775875 +0000 UTC m=+8563.515157154" observedRunningTime="2025-12-12 09:11:16.321147452 +0000 UTC m=+8563.892528761" watchObservedRunningTime="2025-12-12 09:11:16.322532046 +0000 UTC m=+8563.893913315" Dec 12 09:11:58 crc kubenswrapper[4867]: I1212 09:11:58.988997 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:11:58 crc kubenswrapper[4867]: I1212 09:11:58.990022 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:12:28 crc kubenswrapper[4867]: I1212 09:12:28.988963 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:12:28 crc kubenswrapper[4867]: I1212 09:12:28.989699 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:12:58 crc kubenswrapper[4867]: I1212 09:12:58.989453 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:12:58 crc kubenswrapper[4867]: I1212 09:12:58.990521 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:12:58 crc kubenswrapper[4867]: I1212 09:12:58.990626 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:12:58 crc kubenswrapper[4867]: I1212 09:12:58.992033 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:12:58 crc kubenswrapper[4867]: I1212 09:12:58.992142 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445" gracePeriod=600 Dec 12 09:12:59 crc kubenswrapper[4867]: I1212 09:12:59.431201 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445" exitCode=0 Dec 12 09:12:59 crc kubenswrapper[4867]: I1212 09:12:59.431263 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445"} Dec 12 09:12:59 crc kubenswrapper[4867]: I1212 09:12:59.431312 4867 scope.go:117] "RemoveContainer" containerID="225eaab0551e36e117b0b860c85beb16766e1c5c382d60a396b795ed95cfd027" Dec 12 09:13:00 crc kubenswrapper[4867]: I1212 09:13:00.441473 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6"} Dec 12 09:14:21 crc kubenswrapper[4867]: I1212 09:14:21.210315 4867 generic.go:334] "Generic (PLEG): container finished" podID="61a43cf7-2462-4a34-a6f8-4c95a7571df5" containerID="f75e75c901b98c61390c6d22a64209ddab0ea6daf716f0d912708878478120ba" exitCode=0 Dec 12 09:14:21 crc kubenswrapper[4867]: I1212 09:14:21.210399 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" event={"ID":"61a43cf7-2462-4a34-a6f8-4c95a7571df5","Type":"ContainerDied","Data":"f75e75c901b98c61390c6d22a64209ddab0ea6daf716f0d912708878478120ba"} Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.727786 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.761269 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key\") pod \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.761303 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0\") pod \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.761381 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xs9l\" (UniqueName: \"kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l\") pod \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.761494 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle\") pod \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.761553 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory\") pod \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\" (UID: \"61a43cf7-2462-4a34-a6f8-4c95a7571df5\") " Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.768377 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "61a43cf7-2462-4a34-a6f8-4c95a7571df5" (UID: "61a43cf7-2462-4a34-a6f8-4c95a7571df5"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.769515 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l" (OuterVolumeSpecName: "kube-api-access-7xs9l") pod "61a43cf7-2462-4a34-a6f8-4c95a7571df5" (UID: "61a43cf7-2462-4a34-a6f8-4c95a7571df5"). InnerVolumeSpecName "kube-api-access-7xs9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.791615 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory" (OuterVolumeSpecName: "inventory") pod "61a43cf7-2462-4a34-a6f8-4c95a7571df5" (UID: "61a43cf7-2462-4a34-a6f8-4c95a7571df5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.793467 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "61a43cf7-2462-4a34-a6f8-4c95a7571df5" (UID: "61a43cf7-2462-4a34-a6f8-4c95a7571df5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.797919 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "61a43cf7-2462-4a34-a6f8-4c95a7571df5" (UID: "61a43cf7-2462-4a34-a6f8-4c95a7571df5"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.864241 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.864273 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.864285 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xs9l\" (UniqueName: \"kubernetes.io/projected/61a43cf7-2462-4a34-a6f8-4c95a7571df5-kube-api-access-7xs9l\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.864294 4867 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:22 crc kubenswrapper[4867]: I1212 09:14:22.864305 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61a43cf7-2462-4a34-a6f8-4c95a7571df5-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:23 crc kubenswrapper[4867]: I1212 09:14:23.228859 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" event={"ID":"61a43cf7-2462-4a34-a6f8-4c95a7571df5","Type":"ContainerDied","Data":"ea6e37de6b81fc1f02dda50fac57be91fee50882219d86051804c13db584121b"} Dec 12 09:14:23 crc kubenswrapper[4867]: I1212 09:14:23.229379 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea6e37de6b81fc1f02dda50fac57be91fee50882219d86051804c13db584121b" Dec 12 09:14:23 crc kubenswrapper[4867]: I1212 09:14:23.228937 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-87rk6" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.234721 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:14:53 crc kubenswrapper[4867]: E1212 09:14:53.236020 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a43cf7-2462-4a34-a6f8-4c95a7571df5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.236034 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a43cf7-2462-4a34-a6f8-4c95a7571df5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.236317 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="61a43cf7-2462-4a34-a6f8-4c95a7571df5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.237700 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.250929 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.281626 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.281672 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.281735 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns28x\" (UniqueName: \"kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.389679 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.389744 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.389898 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns28x\" (UniqueName: \"kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.391067 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.391388 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.407948 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns28x\" (UniqueName: \"kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x\") pod \"community-operators-57sl4\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:53 crc kubenswrapper[4867]: I1212 09:14:53.570966 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:14:54 crc kubenswrapper[4867]: I1212 09:14:54.107550 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:14:54 crc kubenswrapper[4867]: I1212 09:14:54.511602 4867 generic.go:334] "Generic (PLEG): container finished" podID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerID="0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473" exitCode=0 Dec 12 09:14:54 crc kubenswrapper[4867]: I1212 09:14:54.511683 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerDied","Data":"0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473"} Dec 12 09:14:54 crc kubenswrapper[4867]: I1212 09:14:54.511870 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerStarted","Data":"7c3b9db1b41bd23620e023ca49e3a07ce0d1ffd1a38c9b607539fcb267ee3300"} Dec 12 09:14:54 crc kubenswrapper[4867]: I1212 09:14:54.513524 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.228810 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.229256 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" containerID="cri-o://91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" gracePeriod=30 Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.235660 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.237664 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.238046 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.243036 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.243099 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Liveness" pod="openstack/nova-cell0-conductor-0" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.244259 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.245557 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:55 crc kubenswrapper[4867]: E1212 09:14:55.245605 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.258155 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.258392 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerName="nova-cell1-conductor-conductor" containerID="cri-o://ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" gracePeriod=30 Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.522191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerStarted","Data":"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9"} Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.631016 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.634515 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.647660 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.737491 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.737539 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.737580 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjpdm\" (UniqueName: \"kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.838696 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.838741 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.838773 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjpdm\" (UniqueName: \"kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.839468 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.839693 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.861121 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjpdm\" (UniqueName: \"kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm\") pod \"redhat-marketplace-w6z8d\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:55 crc kubenswrapper[4867]: I1212 09:14:55.955965 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:14:56 crc kubenswrapper[4867]: I1212 09:14:56.531690 4867 generic.go:334] "Generic (PLEG): container finished" podID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerID="32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9" exitCode=0 Dec 12 09:14:56 crc kubenswrapper[4867]: I1212 09:14:56.531783 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerDied","Data":"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9"} Dec 12 09:14:56 crc kubenswrapper[4867]: E1212 09:14:56.556115 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:56 crc kubenswrapper[4867]: E1212 09:14:56.557701 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:56 crc kubenswrapper[4867]: E1212 09:14:56.559411 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 12 09:14:56 crc kubenswrapper[4867]: E1212 09:14:56.559476 4867 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerName="nova-cell1-conductor-conductor" Dec 12 09:14:56 crc kubenswrapper[4867]: I1212 09:14:56.680768 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.007983 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.008292 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" containerID="cri-o://3bb1311ff7539471b027757640ae7cc8d243734703689b5318699c5117b8153f" gracePeriod=30 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.008459 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" containerID="cri-o://6d5a5645c2bac9e863274b960647d85e8f37b212933965ce7e5c348ffaac2620" gracePeriod=30 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.042240 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.042742 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="750d05ab-0045-4a18-9fe8-e9d59566728c" containerName="nova-scheduler-scheduler" containerID="cri-o://94536244d42ee782cd783729152d4af2480ac426bb93d508ffffe5e952fb91ac" gracePeriod=30 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.074651 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.074954 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" containerID="cri-o://6a6a3fad860bb48f1cadf4a3458b5e082f4e5f5c2b4a2f11003c46ac6186158e" gracePeriod=30 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.075091 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" containerID="cri-o://d649cf88d7b70e4ebbf21e46f35b225baed9d1ff65d86e0a97bfeb0016203480" gracePeriod=30 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.545048 4867 generic.go:334] "Generic (PLEG): container finished" podID="1e32b87d-0652-479c-812b-7b93744a3525" containerID="6a6a3fad860bb48f1cadf4a3458b5e082f4e5f5c2b4a2f11003c46ac6186158e" exitCode=143 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.545139 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerDied","Data":"6a6a3fad860bb48f1cadf4a3458b5e082f4e5f5c2b4a2f11003c46ac6186158e"} Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.547597 4867 generic.go:334] "Generic (PLEG): container finished" podID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerID="6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0" exitCode=0 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.547654 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerDied","Data":"6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0"} Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.547671 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerStarted","Data":"a614d219ce2355fbd35bee2c01f955c2925b0f8e0ff1d4006539bdf746877f22"} Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.550286 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerStarted","Data":"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01"} Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.552798 4867 generic.go:334] "Generic (PLEG): container finished" podID="16243454-7b73-4448-b1c6-089261fef413" containerID="3bb1311ff7539471b027757640ae7cc8d243734703689b5318699c5117b8153f" exitCode=143 Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.552901 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerDied","Data":"3bb1311ff7539471b027757640ae7cc8d243734703689b5318699c5117b8153f"} Dec 12 09:14:57 crc kubenswrapper[4867]: I1212 09:14:57.586035 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-57sl4" podStartSLOduration=1.846315752 podStartE2EDuration="4.586011245s" podCreationTimestamp="2025-12-12 09:14:53 +0000 UTC" firstStartedPulling="2025-12-12 09:14:54.513238079 +0000 UTC m=+8782.084619348" lastFinishedPulling="2025-12-12 09:14:57.252933572 +0000 UTC m=+8784.824314841" observedRunningTime="2025-12-12 09:14:57.580869901 +0000 UTC m=+8785.152251170" watchObservedRunningTime="2025-12-12 09:14:57.586011245 +0000 UTC m=+8785.157392514" Dec 12 09:14:58 crc kubenswrapper[4867]: I1212 09:14:58.567482 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerStarted","Data":"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.409340 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.513738 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2896\" (UniqueName: \"kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896\") pod \"d3481dcf-092f-46a3-a41a-81f655801ddb\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.513819 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data\") pod \"d3481dcf-092f-46a3-a41a-81f655801ddb\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.513945 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle\") pod \"d3481dcf-092f-46a3-a41a-81f655801ddb\" (UID: \"d3481dcf-092f-46a3-a41a-81f655801ddb\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.518736 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896" (OuterVolumeSpecName: "kube-api-access-d2896") pod "d3481dcf-092f-46a3-a41a-81f655801ddb" (UID: "d3481dcf-092f-46a3-a41a-81f655801ddb"). InnerVolumeSpecName "kube-api-access-d2896". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.541907 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data" (OuterVolumeSpecName: "config-data") pod "d3481dcf-092f-46a3-a41a-81f655801ddb" (UID: "d3481dcf-092f-46a3-a41a-81f655801ddb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.543411 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3481dcf-092f-46a3-a41a-81f655801ddb" (UID: "d3481dcf-092f-46a3-a41a-81f655801ddb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.581155 4867 generic.go:334] "Generic (PLEG): container finished" podID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerID="6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1" exitCode=0 Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.581297 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerDied","Data":"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.587799 4867 generic.go:334] "Generic (PLEG): container finished" podID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" exitCode=0 Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.587862 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b556b6f3-e3b7-4459-9cd2-86c00b610173","Type":"ContainerDied","Data":"91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.587888 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b556b6f3-e3b7-4459-9cd2-86c00b610173","Type":"ContainerDied","Data":"36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.587899 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36374f5ae3add6a7519c69b48464fca5ba12bf762c1349c19152fc069150b359" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.589520 4867 generic.go:334] "Generic (PLEG): container finished" podID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" exitCode=0 Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.589558 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3481dcf-092f-46a3-a41a-81f655801ddb","Type":"ContainerDied","Data":"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.589576 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d3481dcf-092f-46a3-a41a-81f655801ddb","Type":"ContainerDied","Data":"2b6bad0fb713b173042c0773d6bd1279aa51c1f5d30e64e2834fabf3636f0220"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.589593 4867 scope.go:117] "RemoveContainer" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.589720 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.602523 4867 generic.go:334] "Generic (PLEG): container finished" podID="750d05ab-0045-4a18-9fe8-e9d59566728c" containerID="94536244d42ee782cd783729152d4af2480ac426bb93d508ffffe5e952fb91ac" exitCode=0 Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.602580 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"750d05ab-0045-4a18-9fe8-e9d59566728c","Type":"ContainerDied","Data":"94536244d42ee782cd783729152d4af2480ac426bb93d508ffffe5e952fb91ac"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.602611 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"750d05ab-0045-4a18-9fe8-e9d59566728c","Type":"ContainerDied","Data":"fff8d2d2e8ae39fbd011330ccaa025416fe47ee4e13502ae16581a45c85bfa8f"} Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.602625 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fff8d2d2e8ae39fbd011330ccaa025416fe47ee4e13502ae16581a45c85bfa8f" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.616703 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.616748 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2896\" (UniqueName: \"kubernetes.io/projected/d3481dcf-092f-46a3-a41a-81f655801ddb-kube-api-access-d2896\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.616760 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3481dcf-092f-46a3-a41a-81f655801ddb-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.636698 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.651744 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.653558 4867 scope.go:117] "RemoveContainer" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" Dec 12 09:14:59 crc kubenswrapper[4867]: E1212 09:14:59.653906 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00\": container with ID starting with ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00 not found: ID does not exist" containerID="ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.653950 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00"} err="failed to get container status \"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00\": rpc error: code = NotFound desc = could not find container \"ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00\": container with ID starting with ad47275232e63b2a451864528ab959dbfa7185e47f716076b88891a9c5938c00 not found: ID does not exist" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.669701 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.680019 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.695339 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:14:59 crc kubenswrapper[4867]: E1212 09:14:59.695853 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerName="nova-cell1-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.695886 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerName="nova-cell1-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: E1212 09:14:59.695924 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="750d05ab-0045-4a18-9fe8-e9d59566728c" containerName="nova-scheduler-scheduler" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.695931 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="750d05ab-0045-4a18-9fe8-e9d59566728c" containerName="nova-scheduler-scheduler" Dec 12 09:14:59 crc kubenswrapper[4867]: E1212 09:14:59.695942 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.695948 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.696125 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="750d05ab-0045-4a18-9fe8-e9d59566728c" containerName="nova-scheduler-scheduler" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.696145 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" containerName="nova-cell1-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.696156 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" containerName="nova-cell0-conductor-conductor" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.696867 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.699519 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.730664 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820319 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data\") pod \"750d05ab-0045-4a18-9fe8-e9d59566728c\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820774 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle\") pod \"750d05ab-0045-4a18-9fe8-e9d59566728c\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820810 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9fxp\" (UniqueName: \"kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp\") pod \"750d05ab-0045-4a18-9fe8-e9d59566728c\" (UID: \"750d05ab-0045-4a18-9fe8-e9d59566728c\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820879 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnl55\" (UniqueName: \"kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55\") pod \"b556b6f3-e3b7-4459-9cd2-86c00b610173\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820939 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle\") pod \"b556b6f3-e3b7-4459-9cd2-86c00b610173\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.820961 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data\") pod \"b556b6f3-e3b7-4459-9cd2-86c00b610173\" (UID: \"b556b6f3-e3b7-4459-9cd2-86c00b610173\") " Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.821368 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.821551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.821620 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl54f\" (UniqueName: \"kubernetes.io/projected/5c642a3e-a465-4936-8da1-55a2db56509b-kube-api-access-tl54f\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.825430 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp" (OuterVolumeSpecName: "kube-api-access-x9fxp") pod "750d05ab-0045-4a18-9fe8-e9d59566728c" (UID: "750d05ab-0045-4a18-9fe8-e9d59566728c"). InnerVolumeSpecName "kube-api-access-x9fxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.825729 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55" (OuterVolumeSpecName: "kube-api-access-xnl55") pod "b556b6f3-e3b7-4459-9cd2-86c00b610173" (UID: "b556b6f3-e3b7-4459-9cd2-86c00b610173"). InnerVolumeSpecName "kube-api-access-xnl55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.847986 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "750d05ab-0045-4a18-9fe8-e9d59566728c" (UID: "750d05ab-0045-4a18-9fe8-e9d59566728c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.848532 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data" (OuterVolumeSpecName: "config-data") pod "b556b6f3-e3b7-4459-9cd2-86c00b610173" (UID: "b556b6f3-e3b7-4459-9cd2-86c00b610173"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.848961 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b556b6f3-e3b7-4459-9cd2-86c00b610173" (UID: "b556b6f3-e3b7-4459-9cd2-86c00b610173"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.850168 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data" (OuterVolumeSpecName: "config-data") pod "750d05ab-0045-4a18-9fe8-e9d59566728c" (UID: "750d05ab-0045-4a18-9fe8-e9d59566728c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.923830 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.923927 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.923975 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl54f\" (UniqueName: \"kubernetes.io/projected/5c642a3e-a465-4936-8da1-55a2db56509b-kube-api-access-tl54f\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924074 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924092 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924100 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/750d05ab-0045-4a18-9fe8-e9d59566728c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924112 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9fxp\" (UniqueName: \"kubernetes.io/projected/750d05ab-0045-4a18-9fe8-e9d59566728c-kube-api-access-x9fxp\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924121 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnl55\" (UniqueName: \"kubernetes.io/projected/b556b6f3-e3b7-4459-9cd2-86c00b610173-kube-api-access-xnl55\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.924130 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b556b6f3-e3b7-4459-9cd2-86c00b610173-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.927895 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.929338 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c642a3e-a465-4936-8da1-55a2db56509b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:14:59 crc kubenswrapper[4867]: I1212 09:14:59.939807 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl54f\" (UniqueName: \"kubernetes.io/projected/5c642a3e-a465-4936-8da1-55a2db56509b-kube-api-access-tl54f\") pod \"nova-cell1-conductor-0\" (UID: \"5c642a3e-a465-4936-8da1-55a2db56509b\") " pod="openstack/nova-cell1-conductor-0" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.021451 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.150350 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq"] Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.152071 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.154544 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.154554 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.168494 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq"] Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.208408 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": read tcp 10.217.0.2:59596->10.217.1.91:8775: read: connection reset by peer" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.208616 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": read tcp 10.217.0.2:59598->10.217.1.91:8775: read: connection reset by peer" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.336866 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.336952 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.337170 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8fq8\" (UniqueName: \"kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.430449 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.94:8774/\": read tcp 10.217.0.2:41594->10.217.1.94:8774: read: connection reset by peer" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.431487 4867 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.94:8774/\": read tcp 10.217.0.2:41606->10.217.1.94:8774: read: connection reset by peer" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.440627 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.440693 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.440784 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8fq8\" (UniqueName: \"kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.442268 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.450439 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.461109 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8fq8\" (UniqueName: \"kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8\") pod \"collect-profiles-29425515-wcnfq\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.476935 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.542443 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 12 09:15:00 crc kubenswrapper[4867]: W1212 09:15:00.592002 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c642a3e_a465_4936_8da1_55a2db56509b.slice/crio-b4de20c30fe3a140f59562fc90b5397d7ca83361a80b578b2009faa6fe9e6526 WatchSource:0}: Error finding container b4de20c30fe3a140f59562fc90b5397d7ca83361a80b578b2009faa6fe9e6526: Status 404 returned error can't find the container with id b4de20c30fe3a140f59562fc90b5397d7ca83361a80b578b2009faa6fe9e6526 Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.638631 4867 generic.go:334] "Generic (PLEG): container finished" podID="16243454-7b73-4448-b1c6-089261fef413" containerID="6d5a5645c2bac9e863274b960647d85e8f37b212933965ce7e5c348ffaac2620" exitCode=0 Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.638712 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerDied","Data":"6d5a5645c2bac9e863274b960647d85e8f37b212933965ce7e5c348ffaac2620"} Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.658345 4867 generic.go:334] "Generic (PLEG): container finished" podID="1e32b87d-0652-479c-812b-7b93744a3525" containerID="d649cf88d7b70e4ebbf21e46f35b225baed9d1ff65d86e0a97bfeb0016203480" exitCode=0 Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.658464 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerDied","Data":"d649cf88d7b70e4ebbf21e46f35b225baed9d1ff65d86e0a97bfeb0016203480"} Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.663334 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5c642a3e-a465-4936-8da1-55a2db56509b","Type":"ContainerStarted","Data":"b4de20c30fe3a140f59562fc90b5397d7ca83361a80b578b2009faa6fe9e6526"} Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.667913 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.672977 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerStarted","Data":"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86"} Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.673504 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.724653 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w6z8d" podStartSLOduration=3.067928746 podStartE2EDuration="5.724629669s" podCreationTimestamp="2025-12-12 09:14:55 +0000 UTC" firstStartedPulling="2025-12-12 09:14:57.549396407 +0000 UTC m=+8785.120777676" lastFinishedPulling="2025-12-12 09:15:00.20609733 +0000 UTC m=+8787.777478599" observedRunningTime="2025-12-12 09:15:00.708706487 +0000 UTC m=+8788.280087766" watchObservedRunningTime="2025-12-12 09:15:00.724629669 +0000 UTC m=+8788.296010938" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.825035 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.870584 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3481dcf-092f-46a3-a41a-81f655801ddb" path="/var/lib/kubelet/pods/d3481dcf-092f-46a3-a41a-81f655801ddb/volumes" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.953850 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.964716 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969255 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs\") pod \"1e32b87d-0652-479c-812b-7b93744a3525\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969340 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle\") pod \"1e32b87d-0652-479c-812b-7b93744a3525\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969387 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data\") pod \"1e32b87d-0652-479c-812b-7b93744a3525\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969432 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs\") pod \"1e32b87d-0652-479c-812b-7b93744a3525\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969590 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5mb7\" (UniqueName: \"kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7\") pod \"1e32b87d-0652-479c-812b-7b93744a3525\" (UID: \"1e32b87d-0652-479c-812b-7b93744a3525\") " Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.969918 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs" (OuterVolumeSpecName: "logs") pod "1e32b87d-0652-479c-812b-7b93744a3525" (UID: "1e32b87d-0652-479c-812b-7b93744a3525"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.970353 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e32b87d-0652-479c-812b-7b93744a3525-logs\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.976343 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:15:00 crc kubenswrapper[4867]: I1212 09:15:00.988848 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.002589 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:15:01 crc kubenswrapper[4867]: E1212 09:15:01.003178 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.003194 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" Dec 12 09:15:01 crc kubenswrapper[4867]: E1212 09:15:01.003217 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.003227 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.003449 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-log" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.003470 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e32b87d-0652-479c-812b-7b93744a3525" containerName="nova-metadata-metadata" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.004259 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.009965 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.023570 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.026386 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.031299 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.045414 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.055945 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.066893 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq"] Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.174734 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.174787 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.174809 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.175075 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49bhc\" (UniqueName: \"kubernetes.io/projected/4065a099-8290-4414-bbc5-58c3b4ccae03-kube-api-access-49bhc\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.175390 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-config-data\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.175416 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k22np\" (UniqueName: \"kubernetes.io/projected/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-kube-api-access-k22np\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.276989 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49bhc\" (UniqueName: \"kubernetes.io/projected/4065a099-8290-4414-bbc5-58c3b4ccae03-kube-api-access-49bhc\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.277377 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k22np\" (UniqueName: \"kubernetes.io/projected/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-kube-api-access-k22np\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.277461 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-config-data\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.277583 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.277656 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.277718 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.615607 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7" (OuterVolumeSpecName: "kube-api-access-f5mb7") pod "1e32b87d-0652-479c-812b-7b93744a3525" (UID: "1e32b87d-0652-479c-812b-7b93744a3525"). InnerVolumeSpecName "kube-api-access-f5mb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.617195 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.617200 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-config-data\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.617393 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.618025 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k22np\" (UniqueName: \"kubernetes.io/projected/2c501b05-328a-4dcc-b688-ff4b2c1c90c3-kube-api-access-k22np\") pod \"nova-scheduler-0\" (UID: \"2c501b05-328a-4dcc-b688-ff4b2c1c90c3\") " pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.618829 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4065a099-8290-4414-bbc5-58c3b4ccae03-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.622158 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49bhc\" (UniqueName: \"kubernetes.io/projected/4065a099-8290-4414-bbc5-58c3b4ccae03-kube-api-access-49bhc\") pod \"nova-cell0-conductor-0\" (UID: \"4065a099-8290-4414-bbc5-58c3b4ccae03\") " pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.637831 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 12 09:15:01 crc kubenswrapper[4867]: W1212 09:15:01.645999 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod884ecf9e_075a_4058_9b15_1fb177944369.slice/crio-da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9 WatchSource:0}: Error finding container da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9: Status 404 returned error can't find the container with id da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9 Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.653940 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data" (OuterVolumeSpecName: "config-data") pod "1e32b87d-0652-479c-812b-7b93744a3525" (UID: "1e32b87d-0652-479c-812b-7b93744a3525"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.666037 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.685464 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e32b87d-0652-479c-812b-7b93744a3525" (UID: "1e32b87d-0652-479c-812b-7b93744a3525"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.693591 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.693889 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.693902 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5mb7\" (UniqueName: \"kubernetes.io/projected/1e32b87d-0652-479c-812b-7b93744a3525-kube-api-access-f5mb7\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.705404 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1e32b87d-0652-479c-812b-7b93744a3525" (UID: "1e32b87d-0652-479c-812b-7b93744a3525"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.717952 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" event={"ID":"884ecf9e-075a-4058-9b15-1fb177944369","Type":"ContainerStarted","Data":"da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9"} Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.722947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"16243454-7b73-4448-b1c6-089261fef413","Type":"ContainerDied","Data":"aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320"} Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.723001 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa74f141a91458011a5945f93520e534a31863abe0758295177bca7cf04cd320" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.731167 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.732457 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1e32b87d-0652-479c-812b-7b93744a3525","Type":"ContainerDied","Data":"1ba42ea77700ac3ff526008cc7af34fbf4eb273115c0f27be6499228d69db9a9"} Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.732493 4867 scope.go:117] "RemoveContainer" containerID="d649cf88d7b70e4ebbf21e46f35b225baed9d1ff65d86e0a97bfeb0016203480" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.796707 4867 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e32b87d-0652-479c-812b-7b93744a3525-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.917886 4867 scope.go:117] "RemoveContainer" containerID="6a6a3fad860bb48f1cadf4a3458b5e082f4e5f5c2b4a2f11003c46ac6186158e" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.966570 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 09:15:01 crc kubenswrapper[4867]: I1212 09:15:01.989446 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.013143 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.044470 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: E1212 09:15:02.044855 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.044868 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" Dec 12 09:15:02 crc kubenswrapper[4867]: E1212 09:15:02.044919 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.044927 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.045122 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-log" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.045136 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="16243454-7b73-4448-b1c6-089261fef413" containerName="nova-api-api" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.046172 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.049425 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.049498 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.067085 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116189 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116275 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116540 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116627 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msgdv\" (UniqueName: \"kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116711 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.116786 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs\") pod \"16243454-7b73-4448-b1c6-089261fef413\" (UID: \"16243454-7b73-4448-b1c6-089261fef413\") " Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.130156 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs" (OuterVolumeSpecName: "logs") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.206088 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv" (OuterVolumeSpecName: "kube-api-access-msgdv") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "kube-api-access-msgdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.237698 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-config-data\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.237818 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ptn\" (UniqueName: \"kubernetes.io/projected/97b35086-f8bc-4035-9685-212989ef3b77-kube-api-access-t4ptn\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.237860 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b35086-f8bc-4035-9685-212989ef3b77-logs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.237896 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.238054 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.238136 4867 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16243454-7b73-4448-b1c6-089261fef413-logs\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.238158 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msgdv\" (UniqueName: \"kubernetes.io/projected/16243454-7b73-4448-b1c6-089261fef413-kube-api-access-msgdv\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.255214 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.275409 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data" (OuterVolumeSpecName: "config-data") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.282401 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.291943 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "16243454-7b73-4448-b1c6-089261fef413" (UID: "16243454-7b73-4448-b1c6-089261fef413"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ptn\" (UniqueName: \"kubernetes.io/projected/97b35086-f8bc-4035-9685-212989ef3b77-kube-api-access-t4ptn\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340529 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b35086-f8bc-4035-9685-212989ef3b77-logs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340559 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340658 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340690 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-config-data\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340756 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340767 4867 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340777 4867 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.340785 4867 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16243454-7b73-4448-b1c6-089261fef413-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.341691 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97b35086-f8bc-4035-9685-212989ef3b77-logs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.345522 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.346143 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-config-data\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.347670 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97b35086-f8bc-4035-9685-212989ef3b77-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.361730 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ptn\" (UniqueName: \"kubernetes.io/projected/97b35086-f8bc-4035-9685-212989ef3b77-kube-api-access-t4ptn\") pod \"nova-metadata-0\" (UID: \"97b35086-f8bc-4035-9685-212989ef3b77\") " pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.377151 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.503374 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.513383 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: W1212 09:15:02.520888 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c501b05_328a_4dcc_b688_ff4b2c1c90c3.slice/crio-e34fa0134021c08b8bba1bd6366db290d9658487dd697f2e4a308e1b61093079 WatchSource:0}: Error finding container e34fa0134021c08b8bba1bd6366db290d9658487dd697f2e4a308e1b61093079: Status 404 returned error can't find the container with id e34fa0134021c08b8bba1bd6366db290d9658487dd697f2e4a308e1b61093079 Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.742717 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5c642a3e-a465-4936-8da1-55a2db56509b","Type":"ContainerStarted","Data":"a9c8d249d2b2fbf2a6911188aec46bfa5ac9a6abd20213ab65692aac35490f62"} Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.744658 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.749673 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c501b05-328a-4dcc-b688-ff4b2c1c90c3","Type":"ContainerStarted","Data":"e34fa0134021c08b8bba1bd6366db290d9658487dd697f2e4a308e1b61093079"} Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.755969 4867 generic.go:334] "Generic (PLEG): container finished" podID="884ecf9e-075a-4058-9b15-1fb177944369" containerID="b372927ff6b58c964729165602d4aff5cda95cdc7ae4e1ded790ae901932afbf" exitCode=0 Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.756056 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" event={"ID":"884ecf9e-075a-4058-9b15-1fb177944369","Type":"ContainerDied","Data":"b372927ff6b58c964729165602d4aff5cda95cdc7ae4e1ded790ae901932afbf"} Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.758771 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4065a099-8290-4414-bbc5-58c3b4ccae03","Type":"ContainerStarted","Data":"d98bc4be716cf843db431e3f73d924337830d5dd730ed0da200565f5d8adfdc5"} Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.766699 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.76668002 podStartE2EDuration="3.76668002s" podCreationTimestamp="2025-12-12 09:14:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:15:02.759023677 +0000 UTC m=+8790.330404956" watchObservedRunningTime="2025-12-12 09:15:02.76668002 +0000 UTC m=+8790.338061289" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.768006 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.809996 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.830420 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.863654 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16243454-7b73-4448-b1c6-089261fef413" path="/var/lib/kubelet/pods/16243454-7b73-4448-b1c6-089261fef413/volumes" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.864600 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e32b87d-0652-479c-812b-7b93744a3525" path="/var/lib/kubelet/pods/1e32b87d-0652-479c-812b-7b93744a3525/volumes" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.865573 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="750d05ab-0045-4a18-9fe8-e9d59566728c" path="/var/lib/kubelet/pods/750d05ab-0045-4a18-9fe8-e9d59566728c/volumes" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.869542 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b556b6f3-e3b7-4459-9cd2-86c00b610173" path="/var/lib/kubelet/pods/b556b6f3-e3b7-4459-9cd2-86c00b610173/volumes" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.870284 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.872945 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.872987 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.873112 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.875909 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.876188 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 12 09:15:02 crc kubenswrapper[4867]: I1212 09:15:02.876606 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055176 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055344 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-config-data\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055427 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31e46da1-e8b1-4279-8410-9c32bea377de-logs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055500 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055573 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rm54\" (UniqueName: \"kubernetes.io/projected/31e46da1-e8b1-4279-8410-9c32bea377de-kube-api-access-6rm54\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.055623 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-public-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.156941 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-config-data\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157088 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31e46da1-e8b1-4279-8410-9c32bea377de-logs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157155 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157194 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rm54\" (UniqueName: \"kubernetes.io/projected/31e46da1-e8b1-4279-8410-9c32bea377de-kube-api-access-6rm54\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157268 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-public-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157297 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.157642 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31e46da1-e8b1-4279-8410-9c32bea377de-logs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.218720 4867 scope.go:117] "RemoveContainer" containerID="6d5a5645c2bac9e863274b960647d85e8f37b212933965ce7e5c348ffaac2620" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.244211 4867 scope.go:117] "RemoveContainer" containerID="3bb1311ff7539471b027757640ae7cc8d243734703689b5318699c5117b8153f" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.264394 4867 scope.go:117] "RemoveContainer" containerID="91b7c0f92236b69175d4b77be9e02f89c9655773b85bf5bae56a6d0f3f11252f" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.284758 4867 scope.go:117] "RemoveContainer" containerID="94536244d42ee782cd783729152d4af2480ac426bb93d508ffffe5e952fb91ac" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.503048 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-public-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.503194 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.504540 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-config-data\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.505316 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31e46da1-e8b1-4279-8410-9c32bea377de-internal-tls-certs\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.509599 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rm54\" (UniqueName: \"kubernetes.io/projected/31e46da1-e8b1-4279-8410-9c32bea377de-kube-api-access-6rm54\") pod \"nova-api-0\" (UID: \"31e46da1-e8b1-4279-8410-9c32bea377de\") " pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.572847 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.572882 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.629039 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.787005 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97b35086-f8bc-4035-9685-212989ef3b77","Type":"ContainerStarted","Data":"6bf496b111efed01e1fc3603f6da813fd79564287fa7e48047a26278f673e6ac"} Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.787041 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97b35086-f8bc-4035-9685-212989ef3b77","Type":"ContainerStarted","Data":"8d7d8a5c3a44d856d034aca2f2aae32113e31f367667129d586331ec0f7c4905"} Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.788598 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c501b05-328a-4dcc-b688-ff4b2c1c90c3","Type":"ContainerStarted","Data":"2a84d506bd168e7f9181d4772b4c7147943335816decc0ceb17e9ab4e78edf07"} Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.795228 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4065a099-8290-4414-bbc5-58c3b4ccae03","Type":"ContainerStarted","Data":"410b49904b20554e1884fb315d9d31f028fc5795317ef633d1b198547af10be5"} Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.795275 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.814675 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.817739 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.817720871 podStartE2EDuration="3.817720871s" podCreationTimestamp="2025-12-12 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:15:03.813711554 +0000 UTC m=+8791.385092833" watchObservedRunningTime="2025-12-12 09:15:03.817720871 +0000 UTC m=+8791.389102140" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.845832 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.845812784 podStartE2EDuration="3.845812784s" podCreationTimestamp="2025-12-12 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:15:03.82771715 +0000 UTC m=+8791.399098419" watchObservedRunningTime="2025-12-12 09:15:03.845812784 +0000 UTC m=+8791.417194053" Dec 12 09:15:03 crc kubenswrapper[4867]: I1212 09:15:03.857253 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.117287 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.221831 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.279196 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8fq8\" (UniqueName: \"kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8\") pod \"884ecf9e-075a-4058-9b15-1fb177944369\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.279568 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume\") pod \"884ecf9e-075a-4058-9b15-1fb177944369\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.279847 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume\") pod \"884ecf9e-075a-4058-9b15-1fb177944369\" (UID: \"884ecf9e-075a-4058-9b15-1fb177944369\") " Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.280686 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume" (OuterVolumeSpecName: "config-volume") pod "884ecf9e-075a-4058-9b15-1fb177944369" (UID: "884ecf9e-075a-4058-9b15-1fb177944369"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.285393 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8" (OuterVolumeSpecName: "kube-api-access-n8fq8") pod "884ecf9e-075a-4058-9b15-1fb177944369" (UID: "884ecf9e-075a-4058-9b15-1fb177944369"). InnerVolumeSpecName "kube-api-access-n8fq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.286403 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "884ecf9e-075a-4058-9b15-1fb177944369" (UID: "884ecf9e-075a-4058-9b15-1fb177944369"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.311160 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.382669 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/884ecf9e-075a-4058-9b15-1fb177944369-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.383074 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/884ecf9e-075a-4058-9b15-1fb177944369-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.383167 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8fq8\" (UniqueName: \"kubernetes.io/projected/884ecf9e-075a-4058-9b15-1fb177944369-kube-api-access-n8fq8\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.804406 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" event={"ID":"884ecf9e-075a-4058-9b15-1fb177944369","Type":"ContainerDied","Data":"da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9"} Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.805940 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da55dab666b15cf0694375db2bd801241c641f42b87398c0407972760a533fd9" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.804439 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425515-wcnfq" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.806191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31e46da1-e8b1-4279-8410-9c32bea377de","Type":"ContainerStarted","Data":"40d0a54e31515db14372e01ea619a4ead9c4158a4b8c8c0cc313577ba233bfbe"} Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.806215 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31e46da1-e8b1-4279-8410-9c32bea377de","Type":"ContainerStarted","Data":"b1995de344dbd7fc9d0093c98092fe0645b6143bd5898cff875a0c4c9725a85d"} Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.806236 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31e46da1-e8b1-4279-8410-9c32bea377de","Type":"ContainerStarted","Data":"42e5dfacfc4cf840a4f7f519642d2bac27dcc8ba8e850e202e92b2cc8c381494"} Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.808609 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97b35086-f8bc-4035-9685-212989ef3b77","Type":"ContainerStarted","Data":"a18bd4b491c2c77d0f2d24e61484b7d9d86be21c9a496f59136b3808e7ea8039"} Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.834771 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.834754285 podStartE2EDuration="2.834754285s" podCreationTimestamp="2025-12-12 09:15:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:15:04.831636071 +0000 UTC m=+8792.403017340" watchObservedRunningTime="2025-12-12 09:15:04.834754285 +0000 UTC m=+8792.406135544" Dec 12 09:15:04 crc kubenswrapper[4867]: I1212 09:15:04.866020 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.865995604 podStartE2EDuration="3.865995604s" podCreationTimestamp="2025-12-12 09:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:15:04.856854425 +0000 UTC m=+8792.428235694" watchObservedRunningTime="2025-12-12 09:15:04.865995604 +0000 UTC m=+8792.437376873" Dec 12 09:15:05 crc kubenswrapper[4867]: I1212 09:15:05.195165 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff"] Dec 12 09:15:05 crc kubenswrapper[4867]: I1212 09:15:05.208113 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425470-dzdff"] Dec 12 09:15:05 crc kubenswrapper[4867]: I1212 09:15:05.817113 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-57sl4" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="registry-server" containerID="cri-o://873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01" gracePeriod=2 Dec 12 09:15:05 crc kubenswrapper[4867]: I1212 09:15:05.956608 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:05 crc kubenswrapper[4867]: I1212 09:15:05.957878 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.009907 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.368797 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.435405 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities\") pod \"ea4a0875-0072-4b55-89d7-52ccff38d252\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.435557 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content\") pod \"ea4a0875-0072-4b55-89d7-52ccff38d252\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.435653 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns28x\" (UniqueName: \"kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x\") pod \"ea4a0875-0072-4b55-89d7-52ccff38d252\" (UID: \"ea4a0875-0072-4b55-89d7-52ccff38d252\") " Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.436993 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities" (OuterVolumeSpecName: "utilities") pod "ea4a0875-0072-4b55-89d7-52ccff38d252" (UID: "ea4a0875-0072-4b55-89d7-52ccff38d252"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.446628 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x" (OuterVolumeSpecName: "kube-api-access-ns28x") pod "ea4a0875-0072-4b55-89d7-52ccff38d252" (UID: "ea4a0875-0072-4b55-89d7-52ccff38d252"). InnerVolumeSpecName "kube-api-access-ns28x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.490206 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea4a0875-0072-4b55-89d7-52ccff38d252" (UID: "ea4a0875-0072-4b55-89d7-52ccff38d252"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.538115 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.538154 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea4a0875-0072-4b55-89d7-52ccff38d252-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.538171 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns28x\" (UniqueName: \"kubernetes.io/projected/ea4a0875-0072-4b55-89d7-52ccff38d252-kube-api-access-ns28x\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.638610 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.827687 4867 generic.go:334] "Generic (PLEG): container finished" podID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerID="873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01" exitCode=0 Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.827743 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerDied","Data":"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01"} Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.827786 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57sl4" event={"ID":"ea4a0875-0072-4b55-89d7-52ccff38d252","Type":"ContainerDied","Data":"7c3b9db1b41bd23620e023ca49e3a07ce0d1ffd1a38c9b607539fcb267ee3300"} Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.827805 4867 scope.go:117] "RemoveContainer" containerID="873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.827758 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57sl4" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.850082 4867 scope.go:117] "RemoveContainer" containerID="32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.874772 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f1ab72-a940-4d62-a67c-367662591295" path="/var/lib/kubelet/pods/83f1ab72-a940-4d62-a67c-367662591295/volumes" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.876035 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.879463 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-57sl4"] Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.892023 4867 scope.go:117] "RemoveContainer" containerID="0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.896831 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.946200 4867 scope.go:117] "RemoveContainer" containerID="873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01" Dec 12 09:15:06 crc kubenswrapper[4867]: E1212 09:15:06.946948 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01\": container with ID starting with 873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01 not found: ID does not exist" containerID="873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.946977 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01"} err="failed to get container status \"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01\": rpc error: code = NotFound desc = could not find container \"873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01\": container with ID starting with 873f71f739c3a6be33ad70b533efa67d713845262bbdbf37157371afb899ef01 not found: ID does not exist" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.946997 4867 scope.go:117] "RemoveContainer" containerID="32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9" Dec 12 09:15:06 crc kubenswrapper[4867]: E1212 09:15:06.947835 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9\": container with ID starting with 32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9 not found: ID does not exist" containerID="32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.947880 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9"} err="failed to get container status \"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9\": rpc error: code = NotFound desc = could not find container \"32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9\": container with ID starting with 32fb9dbb631231c49e826b2a68618b487d124314f390241d3aadde174015f2e9 not found: ID does not exist" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.947908 4867 scope.go:117] "RemoveContainer" containerID="0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473" Dec 12 09:15:06 crc kubenswrapper[4867]: E1212 09:15:06.948298 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473\": container with ID starting with 0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473 not found: ID does not exist" containerID="0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473" Dec 12 09:15:06 crc kubenswrapper[4867]: I1212 09:15:06.948330 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473"} err="failed to get container status \"0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473\": rpc error: code = NotFound desc = could not find container \"0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473\": container with ID starting with 0b8c203d1d37cea752966c69e8e877ee03d16b3461e0cfe7df63f81d12528473 not found: ID does not exist" Dec 12 09:15:07 crc kubenswrapper[4867]: I1212 09:15:07.377982 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 09:15:07 crc kubenswrapper[4867]: I1212 09:15:07.378157 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 12 09:15:08 crc kubenswrapper[4867]: I1212 09:15:08.420377 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:15:08 crc kubenswrapper[4867]: I1212 09:15:08.849200 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" path="/var/lib/kubelet/pods/ea4a0875-0072-4b55-89d7-52ccff38d252/volumes" Dec 12 09:15:09 crc kubenswrapper[4867]: I1212 09:15:09.862066 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w6z8d" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="registry-server" containerID="cri-o://f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86" gracePeriod=2 Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.054619 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.354106 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.522944 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content\") pod \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.523133 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities\") pod \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.523271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjpdm\" (UniqueName: \"kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm\") pod \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\" (UID: \"1bf0ec8a-409e-49c7-8700-c21c5bc94d59\") " Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.524124 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities" (OuterVolumeSpecName: "utilities") pod "1bf0ec8a-409e-49c7-8700-c21c5bc94d59" (UID: "1bf0ec8a-409e-49c7-8700-c21c5bc94d59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.529285 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm" (OuterVolumeSpecName: "kube-api-access-qjpdm") pod "1bf0ec8a-409e-49c7-8700-c21c5bc94d59" (UID: "1bf0ec8a-409e-49c7-8700-c21c5bc94d59"). InnerVolumeSpecName "kube-api-access-qjpdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.545114 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bf0ec8a-409e-49c7-8700-c21c5bc94d59" (UID: "1bf0ec8a-409e-49c7-8700-c21c5bc94d59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.625656 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.625690 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjpdm\" (UniqueName: \"kubernetes.io/projected/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-kube-api-access-qjpdm\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.625699 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bf0ec8a-409e-49c7-8700-c21c5bc94d59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.873739 4867 generic.go:334] "Generic (PLEG): container finished" podID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerID="f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86" exitCode=0 Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.873787 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerDied","Data":"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86"} Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.873818 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w6z8d" event={"ID":"1bf0ec8a-409e-49c7-8700-c21c5bc94d59","Type":"ContainerDied","Data":"a614d219ce2355fbd35bee2c01f955c2925b0f8e0ff1d4006539bdf746877f22"} Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.873838 4867 scope.go:117] "RemoveContainer" containerID="f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.873989 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w6z8d" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.902904 4867 scope.go:117] "RemoveContainer" containerID="6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.903900 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.912851 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w6z8d"] Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.926134 4867 scope.go:117] "RemoveContainer" containerID="6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.972867 4867 scope.go:117] "RemoveContainer" containerID="f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86" Dec 12 09:15:10 crc kubenswrapper[4867]: E1212 09:15:10.973736 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86\": container with ID starting with f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86 not found: ID does not exist" containerID="f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.973809 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86"} err="failed to get container status \"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86\": rpc error: code = NotFound desc = could not find container \"f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86\": container with ID starting with f151e86ca7d879ec453b8d9a784d6df7d84b6d8dc7698a478529b6a6fc9bdd86 not found: ID does not exist" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.973839 4867 scope.go:117] "RemoveContainer" containerID="6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1" Dec 12 09:15:10 crc kubenswrapper[4867]: E1212 09:15:10.974391 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1\": container with ID starting with 6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1 not found: ID does not exist" containerID="6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.974445 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1"} err="failed to get container status \"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1\": rpc error: code = NotFound desc = could not find container \"6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1\": container with ID starting with 6d6f69d80d0e820f331bd02c8801f0109d707ec61649e037076424bc26c719b1 not found: ID does not exist" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.974478 4867 scope.go:117] "RemoveContainer" containerID="6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0" Dec 12 09:15:10 crc kubenswrapper[4867]: E1212 09:15:10.975197 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0\": container with ID starting with 6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0 not found: ID does not exist" containerID="6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0" Dec 12 09:15:10 crc kubenswrapper[4867]: I1212 09:15:10.975223 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0"} err="failed to get container status \"6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0\": rpc error: code = NotFound desc = could not find container \"6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0\": container with ID starting with 6cb9e67cca8c716c5a0c567c9a05ae179e64171d73cd136a9d8b3bf7d5fe8ea0 not found: ID does not exist" Dec 12 09:15:11 crc kubenswrapper[4867]: I1212 09:15:11.639174 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 12 09:15:11 crc kubenswrapper[4867]: I1212 09:15:11.673087 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 12 09:15:11 crc kubenswrapper[4867]: I1212 09:15:11.698989 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 12 09:15:11 crc kubenswrapper[4867]: I1212 09:15:11.913347 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 12 09:15:12 crc kubenswrapper[4867]: I1212 09:15:12.377679 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 09:15:12 crc kubenswrapper[4867]: I1212 09:15:12.377748 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 12 09:15:12 crc kubenswrapper[4867]: I1212 09:15:12.849701 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" path="/var/lib/kubelet/pods/1bf0ec8a-409e-49c7-8700-c21c5bc94d59/volumes" Dec 12 09:15:13 crc kubenswrapper[4867]: I1212 09:15:13.390410 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="97b35086-f8bc-4035-9685-212989ef3b77" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 09:15:13 crc kubenswrapper[4867]: I1212 09:15:13.390430 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="97b35086-f8bc-4035-9685-212989ef3b77" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 09:15:13 crc kubenswrapper[4867]: I1212 09:15:13.815406 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 09:15:13 crc kubenswrapper[4867]: I1212 09:15:13.815448 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 12 09:15:14 crc kubenswrapper[4867]: I1212 09:15:14.829418 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31e46da1-e8b1-4279-8410-9c32bea377de" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 09:15:14 crc kubenswrapper[4867]: I1212 09:15:14.829473 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31e46da1-e8b1-4279-8410-9c32bea377de" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 12 09:15:22 crc kubenswrapper[4867]: I1212 09:15:22.383091 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 09:15:22 crc kubenswrapper[4867]: I1212 09:15:22.383651 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 12 09:15:22 crc kubenswrapper[4867]: I1212 09:15:22.390326 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 09:15:22 crc kubenswrapper[4867]: I1212 09:15:22.391672 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.823079 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.823416 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.823632 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.823659 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.830315 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 09:15:23 crc kubenswrapper[4867]: I1212 09:15:23.831739 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849300 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d"] Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849634 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849646 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849666 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="extract-utilities" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849672 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="extract-utilities" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849688 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="extract-utilities" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849694 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="extract-utilities" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849711 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="extract-content" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849717 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="extract-content" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849726 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="extract-content" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849733 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="extract-content" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849748 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="884ecf9e-075a-4058-9b15-1fb177944369" containerName="collect-profiles" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849753 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="884ecf9e-075a-4058-9b15-1fb177944369" containerName="collect-profiles" Dec 12 09:15:24 crc kubenswrapper[4867]: E1212 09:15:24.849760 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.849766 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.850013 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea4a0875-0072-4b55-89d7-52ccff38d252" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.850033 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf0ec8a-409e-49c7-8700-c21c5bc94d59" containerName="registry-server" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.850047 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="884ecf9e-075a-4058-9b15-1fb177944369" containerName="collect-profiles" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.850724 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.853911 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.854539 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.854585 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.855097 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.855315 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-smvch" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.855351 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.855328 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.870422 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d"] Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.999258 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.999551 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.999652 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ml49\" (UniqueName: \"kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.999727 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:24 crc kubenswrapper[4867]: I1212 09:15:24.999899 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:24.999998 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.000164 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.000269 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.000383 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102313 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102390 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102436 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ml49\" (UniqueName: \"kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102471 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102536 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102574 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102663 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102695 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.102738 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.104543 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.108481 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.114998 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.115194 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.115333 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.115423 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.115438 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.115438 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.126240 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ml49\" (UniqueName: \"kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.172176 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:15:25 crc kubenswrapper[4867]: I1212 09:15:25.745607 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d"] Dec 12 09:15:26 crc kubenswrapper[4867]: I1212 09:15:26.017168 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" event={"ID":"7307958d-09aa-4a51-b62b-7dac152de62a","Type":"ContainerStarted","Data":"03f5dede87006755df0443b86a8bc091144c2f1865aeee76ca68bd7388ecb0aa"} Dec 12 09:15:27 crc kubenswrapper[4867]: I1212 09:15:27.027378 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" event={"ID":"7307958d-09aa-4a51-b62b-7dac152de62a","Type":"ContainerStarted","Data":"1cd3002cc7cbb387954e1a1baa6fef0c0704cfadbad74bb4f351f832285e76f3"} Dec 12 09:15:27 crc kubenswrapper[4867]: I1212 09:15:27.050382 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" podStartSLOduration=2.397307583 podStartE2EDuration="3.050343754s" podCreationTimestamp="2025-12-12 09:15:24 +0000 UTC" firstStartedPulling="2025-12-12 09:15:25.755378677 +0000 UTC m=+8813.326759946" lastFinishedPulling="2025-12-12 09:15:26.408414848 +0000 UTC m=+8813.979796117" observedRunningTime="2025-12-12 09:15:27.04439085 +0000 UTC m=+8814.615772119" watchObservedRunningTime="2025-12-12 09:15:27.050343754 +0000 UTC m=+8814.621725023" Dec 12 09:15:28 crc kubenswrapper[4867]: I1212 09:15:28.990623 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:15:28 crc kubenswrapper[4867]: I1212 09:15:28.990908 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:15:58 crc kubenswrapper[4867]: I1212 09:15:58.988663 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:15:58 crc kubenswrapper[4867]: I1212 09:15:58.989303 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:16:03 crc kubenswrapper[4867]: I1212 09:16:03.741657 4867 scope.go:117] "RemoveContainer" containerID="99b5596959fc5842036d24edfe78ffef6eed4e790cead83e4dba8aa177600f84" Dec 12 09:16:28 crc kubenswrapper[4867]: I1212 09:16:28.989508 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:16:28 crc kubenswrapper[4867]: I1212 09:16:28.989995 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:16:28 crc kubenswrapper[4867]: I1212 09:16:28.990035 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:16:28 crc kubenswrapper[4867]: I1212 09:16:28.990805 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:16:28 crc kubenswrapper[4867]: I1212 09:16:28.990855 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" gracePeriod=600 Dec 12 09:16:29 crc kubenswrapper[4867]: E1212 09:16:29.163578 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:16:29 crc kubenswrapper[4867]: I1212 09:16:29.613434 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" exitCode=0 Dec 12 09:16:29 crc kubenswrapper[4867]: I1212 09:16:29.613468 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6"} Dec 12 09:16:29 crc kubenswrapper[4867]: I1212 09:16:29.613531 4867 scope.go:117] "RemoveContainer" containerID="c3ea851a08406aebefb3ea227e9c2589e332f38eee7d924547b711f19dfd4445" Dec 12 09:16:29 crc kubenswrapper[4867]: I1212 09:16:29.614280 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:16:29 crc kubenswrapper[4867]: E1212 09:16:29.614563 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:16:44 crc kubenswrapper[4867]: I1212 09:16:44.838633 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:16:44 crc kubenswrapper[4867]: E1212 09:16:44.839513 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:16:58 crc kubenswrapper[4867]: I1212 09:16:58.838085 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:16:58 crc kubenswrapper[4867]: E1212 09:16:58.838896 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:17:09 crc kubenswrapper[4867]: I1212 09:17:09.839324 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:17:09 crc kubenswrapper[4867]: E1212 09:17:09.840027 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:17:20 crc kubenswrapper[4867]: I1212 09:17:20.838584 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:17:20 crc kubenswrapper[4867]: E1212 09:17:20.839697 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:17:32 crc kubenswrapper[4867]: I1212 09:17:32.851844 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:17:32 crc kubenswrapper[4867]: E1212 09:17:32.852831 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:17:47 crc kubenswrapper[4867]: I1212 09:17:47.839330 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:17:47 crc kubenswrapper[4867]: E1212 09:17:47.840150 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:18:00 crc kubenswrapper[4867]: I1212 09:18:00.838980 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:18:00 crc kubenswrapper[4867]: E1212 09:18:00.840066 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:18:12 crc kubenswrapper[4867]: I1212 09:18:12.845759 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:18:12 crc kubenswrapper[4867]: E1212 09:18:12.846628 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:18:24 crc kubenswrapper[4867]: I1212 09:18:24.838983 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:18:24 crc kubenswrapper[4867]: E1212 09:18:24.839736 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:18:38 crc kubenswrapper[4867]: I1212 09:18:38.838600 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:18:38 crc kubenswrapper[4867]: E1212 09:18:38.839395 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:18:51 crc kubenswrapper[4867]: I1212 09:18:51.837866 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:18:51 crc kubenswrapper[4867]: E1212 09:18:51.839453 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:03 crc kubenswrapper[4867]: I1212 09:19:03.838691 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:19:03 crc kubenswrapper[4867]: E1212 09:19:03.839435 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:15 crc kubenswrapper[4867]: I1212 09:19:15.838622 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:19:15 crc kubenswrapper[4867]: E1212 09:19:15.841301 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:27 crc kubenswrapper[4867]: I1212 09:19:27.837839 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:19:27 crc kubenswrapper[4867]: E1212 09:19:27.838874 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:41 crc kubenswrapper[4867]: I1212 09:19:41.839120 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:19:41 crc kubenswrapper[4867]: E1212 09:19:41.840079 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.124015 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.126799 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.156410 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.273367 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tghz\" (UniqueName: \"kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.273472 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.273505 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.375454 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.375513 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.375665 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tghz\" (UniqueName: \"kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.376078 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.376111 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.393690 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tghz\" (UniqueName: \"kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz\") pod \"certified-operators-ht4f5\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.454297 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:45 crc kubenswrapper[4867]: I1212 09:19:45.971498 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:46 crc kubenswrapper[4867]: I1212 09:19:46.521523 4867 generic.go:334] "Generic (PLEG): container finished" podID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerID="e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b" exitCode=0 Dec 12 09:19:46 crc kubenswrapper[4867]: I1212 09:19:46.521582 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerDied","Data":"e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b"} Dec 12 09:19:46 crc kubenswrapper[4867]: I1212 09:19:46.521781 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerStarted","Data":"3aec465431af91471a11f943180498c9abb51538d4c6a7b2fd50530688c8f60e"} Dec 12 09:19:48 crc kubenswrapper[4867]: I1212 09:19:48.541671 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerStarted","Data":"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa"} Dec 12 09:19:49 crc kubenswrapper[4867]: I1212 09:19:49.553137 4867 generic.go:334] "Generic (PLEG): container finished" podID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerID="88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa" exitCode=0 Dec 12 09:19:49 crc kubenswrapper[4867]: I1212 09:19:49.553238 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerDied","Data":"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa"} Dec 12 09:19:51 crc kubenswrapper[4867]: I1212 09:19:51.582352 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerStarted","Data":"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9"} Dec 12 09:19:51 crc kubenswrapper[4867]: I1212 09:19:51.617349 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ht4f5" podStartSLOduration=3.148785581 podStartE2EDuration="6.61731827s" podCreationTimestamp="2025-12-12 09:19:45 +0000 UTC" firstStartedPulling="2025-12-12 09:19:46.524768318 +0000 UTC m=+9074.096149587" lastFinishedPulling="2025-12-12 09:19:49.993301007 +0000 UTC m=+9077.564682276" observedRunningTime="2025-12-12 09:19:51.605840635 +0000 UTC m=+9079.177221904" watchObservedRunningTime="2025-12-12 09:19:51.61731827 +0000 UTC m=+9079.188699559" Dec 12 09:19:53 crc kubenswrapper[4867]: I1212 09:19:53.837837 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:19:53 crc kubenswrapper[4867]: E1212 09:19:53.838497 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:19:55 crc kubenswrapper[4867]: I1212 09:19:55.455150 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:55 crc kubenswrapper[4867]: I1212 09:19:55.455485 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:55 crc kubenswrapper[4867]: I1212 09:19:55.499795 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:55 crc kubenswrapper[4867]: I1212 09:19:55.659237 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:55 crc kubenswrapper[4867]: I1212 09:19:55.740644 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:57 crc kubenswrapper[4867]: I1212 09:19:57.635309 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ht4f5" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="registry-server" containerID="cri-o://e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9" gracePeriod=2 Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.085270 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.239091 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content\") pod \"cbe906f4-a8a4-480e-86da-307842a5fd52\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.239216 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tghz\" (UniqueName: \"kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz\") pod \"cbe906f4-a8a4-480e-86da-307842a5fd52\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.239700 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities\") pod \"cbe906f4-a8a4-480e-86da-307842a5fd52\" (UID: \"cbe906f4-a8a4-480e-86da-307842a5fd52\") " Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.240286 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities" (OuterVolumeSpecName: "utilities") pod "cbe906f4-a8a4-480e-86da-307842a5fd52" (UID: "cbe906f4-a8a4-480e-86da-307842a5fd52"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.245127 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz" (OuterVolumeSpecName: "kube-api-access-2tghz") pod "cbe906f4-a8a4-480e-86da-307842a5fd52" (UID: "cbe906f4-a8a4-480e-86da-307842a5fd52"). InnerVolumeSpecName "kube-api-access-2tghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.304401 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbe906f4-a8a4-480e-86da-307842a5fd52" (UID: "cbe906f4-a8a4-480e-86da-307842a5fd52"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.342038 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.342067 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbe906f4-a8a4-480e-86da-307842a5fd52-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.342079 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tghz\" (UniqueName: \"kubernetes.io/projected/cbe906f4-a8a4-480e-86da-307842a5fd52-kube-api-access-2tghz\") on node \"crc\" DevicePath \"\"" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.646126 4867 generic.go:334] "Generic (PLEG): container finished" podID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerID="e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9" exitCode=0 Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.646172 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerDied","Data":"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9"} Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.646199 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ht4f5" event={"ID":"cbe906f4-a8a4-480e-86da-307842a5fd52","Type":"ContainerDied","Data":"3aec465431af91471a11f943180498c9abb51538d4c6a7b2fd50530688c8f60e"} Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.646216 4867 scope.go:117] "RemoveContainer" containerID="e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.646395 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ht4f5" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.687882 4867 scope.go:117] "RemoveContainer" containerID="88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.691368 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.703474 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ht4f5"] Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.715453 4867 scope.go:117] "RemoveContainer" containerID="e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.760478 4867 scope.go:117] "RemoveContainer" containerID="e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9" Dec 12 09:19:58 crc kubenswrapper[4867]: E1212 09:19:58.761301 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9\": container with ID starting with e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9 not found: ID does not exist" containerID="e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.761354 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9"} err="failed to get container status \"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9\": rpc error: code = NotFound desc = could not find container \"e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9\": container with ID starting with e398c6a87a1316b84abb765679a267c6200beb14f280c8d592d9e17a86072fd9 not found: ID does not exist" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.761383 4867 scope.go:117] "RemoveContainer" containerID="88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa" Dec 12 09:19:58 crc kubenswrapper[4867]: E1212 09:19:58.761859 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa\": container with ID starting with 88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa not found: ID does not exist" containerID="88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.761891 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa"} err="failed to get container status \"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa\": rpc error: code = NotFound desc = could not find container \"88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa\": container with ID starting with 88a36a0c153966d5e27fd8003b9c06d705f18ded6e4b25abd0fdfed4ab207efa not found: ID does not exist" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.761915 4867 scope.go:117] "RemoveContainer" containerID="e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b" Dec 12 09:19:58 crc kubenswrapper[4867]: E1212 09:19:58.762208 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b\": container with ID starting with e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b not found: ID does not exist" containerID="e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.762244 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b"} err="failed to get container status \"e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b\": rpc error: code = NotFound desc = could not find container \"e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b\": container with ID starting with e118aa33c203bfe7f45267400d99a11426e2af1f517aa2d841915e68e2292d2b not found: ID does not exist" Dec 12 09:19:58 crc kubenswrapper[4867]: I1212 09:19:58.849957 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" path="/var/lib/kubelet/pods/cbe906f4-a8a4-480e-86da-307842a5fd52/volumes" Dec 12 09:20:06 crc kubenswrapper[4867]: I1212 09:20:06.838526 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:20:06 crc kubenswrapper[4867]: E1212 09:20:06.839415 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:20:20 crc kubenswrapper[4867]: I1212 09:20:20.838287 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:20:20 crc kubenswrapper[4867]: E1212 09:20:20.839143 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:20:35 crc kubenswrapper[4867]: I1212 09:20:35.838424 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:20:35 crc kubenswrapper[4867]: E1212 09:20:35.839868 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:20:48 crc kubenswrapper[4867]: I1212 09:20:48.839429 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:20:48 crc kubenswrapper[4867]: E1212 09:20:48.840453 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.936977 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:20:53 crc kubenswrapper[4867]: E1212 09:20:53.937927 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="registry-server" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.937939 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="registry-server" Dec 12 09:20:53 crc kubenswrapper[4867]: E1212 09:20:53.937955 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="extract-content" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.937961 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="extract-content" Dec 12 09:20:53 crc kubenswrapper[4867]: E1212 09:20:53.937981 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="extract-utilities" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.937988 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="extract-utilities" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.938204 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe906f4-a8a4-480e-86da-307842a5fd52" containerName="registry-server" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.941063 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:53 crc kubenswrapper[4867]: I1212 09:20:53.950995 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.057849 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.057926 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4pxl\" (UniqueName: \"kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.058069 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.159546 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.159897 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.160007 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4pxl\" (UniqueName: \"kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.160033 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.160258 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.179426 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4pxl\" (UniqueName: \"kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl\") pod \"redhat-operators-bnwhf\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.266544 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:20:54 crc kubenswrapper[4867]: I1212 09:20:54.792153 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:20:56 crc kubenswrapper[4867]: I1212 09:20:56.187034 4867 generic.go:334] "Generic (PLEG): container finished" podID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerID="589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125" exitCode=0 Dec 12 09:20:56 crc kubenswrapper[4867]: I1212 09:20:56.187097 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerDied","Data":"589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125"} Dec 12 09:20:56 crc kubenswrapper[4867]: I1212 09:20:56.187493 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerStarted","Data":"bef891b7a41828466bd8a3cc9d3b2495db5c7272b045c2a0696257bbb63bc8a7"} Dec 12 09:20:56 crc kubenswrapper[4867]: I1212 09:20:56.191032 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:20:58 crc kubenswrapper[4867]: I1212 09:20:58.214057 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerStarted","Data":"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9"} Dec 12 09:21:02 crc kubenswrapper[4867]: I1212 09:21:02.283701 4867 generic.go:334] "Generic (PLEG): container finished" podID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerID="6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9" exitCode=0 Dec 12 09:21:02 crc kubenswrapper[4867]: I1212 09:21:02.283750 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerDied","Data":"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9"} Dec 12 09:21:02 crc kubenswrapper[4867]: I1212 09:21:02.847639 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:21:02 crc kubenswrapper[4867]: E1212 09:21:02.848205 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:21:03 crc kubenswrapper[4867]: I1212 09:21:03.294003 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerStarted","Data":"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f"} Dec 12 09:21:03 crc kubenswrapper[4867]: I1212 09:21:03.329322 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bnwhf" podStartSLOduration=3.578115032 podStartE2EDuration="10.329295583s" podCreationTimestamp="2025-12-12 09:20:53 +0000 UTC" firstStartedPulling="2025-12-12 09:20:56.19057791 +0000 UTC m=+9143.761959189" lastFinishedPulling="2025-12-12 09:21:02.941758471 +0000 UTC m=+9150.513139740" observedRunningTime="2025-12-12 09:21:03.314462376 +0000 UTC m=+9150.885843645" watchObservedRunningTime="2025-12-12 09:21:03.329295583 +0000 UTC m=+9150.900676852" Dec 12 09:21:04 crc kubenswrapper[4867]: I1212 09:21:04.267856 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:04 crc kubenswrapper[4867]: I1212 09:21:04.268786 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:05 crc kubenswrapper[4867]: I1212 09:21:05.313332 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bnwhf" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="registry-server" probeResult="failure" output=< Dec 12 09:21:05 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 09:21:05 crc kubenswrapper[4867]: > Dec 12 09:21:14 crc kubenswrapper[4867]: I1212 09:21:14.313816 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:14 crc kubenswrapper[4867]: I1212 09:21:14.360890 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:14 crc kubenswrapper[4867]: I1212 09:21:14.560374 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:21:15 crc kubenswrapper[4867]: I1212 09:21:15.421420 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bnwhf" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="registry-server" containerID="cri-o://0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f" gracePeriod=2 Dec 12 09:21:15 crc kubenswrapper[4867]: I1212 09:21:15.837975 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:21:15 crc kubenswrapper[4867]: E1212 09:21:15.838301 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:21:15 crc kubenswrapper[4867]: I1212 09:21:15.878003 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.028153 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities\") pod \"709ec057-02f2-48ce-88ce-73c37d8d8a44\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.028490 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4pxl\" (UniqueName: \"kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl\") pod \"709ec057-02f2-48ce-88ce-73c37d8d8a44\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.028583 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content\") pod \"709ec057-02f2-48ce-88ce-73c37d8d8a44\" (UID: \"709ec057-02f2-48ce-88ce-73c37d8d8a44\") " Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.029165 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities" (OuterVolumeSpecName: "utilities") pod "709ec057-02f2-48ce-88ce-73c37d8d8a44" (UID: "709ec057-02f2-48ce-88ce-73c37d8d8a44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.036245 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl" (OuterVolumeSpecName: "kube-api-access-s4pxl") pod "709ec057-02f2-48ce-88ce-73c37d8d8a44" (UID: "709ec057-02f2-48ce-88ce-73c37d8d8a44"). InnerVolumeSpecName "kube-api-access-s4pxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.132759 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.132820 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4pxl\" (UniqueName: \"kubernetes.io/projected/709ec057-02f2-48ce-88ce-73c37d8d8a44-kube-api-access-s4pxl\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.149439 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "709ec057-02f2-48ce-88ce-73c37d8d8a44" (UID: "709ec057-02f2-48ce-88ce-73c37d8d8a44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.234576 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/709ec057-02f2-48ce-88ce-73c37d8d8a44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.433413 4867 generic.go:334] "Generic (PLEG): container finished" podID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerID="0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f" exitCode=0 Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.433493 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerDied","Data":"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f"} Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.433538 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnwhf" event={"ID":"709ec057-02f2-48ce-88ce-73c37d8d8a44","Type":"ContainerDied","Data":"bef891b7a41828466bd8a3cc9d3b2495db5c7272b045c2a0696257bbb63bc8a7"} Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.433563 4867 scope.go:117] "RemoveContainer" containerID="0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.433818 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnwhf" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.457370 4867 scope.go:117] "RemoveContainer" containerID="6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.477474 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.487113 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bnwhf"] Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.510762 4867 scope.go:117] "RemoveContainer" containerID="589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.534440 4867 scope.go:117] "RemoveContainer" containerID="0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f" Dec 12 09:21:16 crc kubenswrapper[4867]: E1212 09:21:16.535079 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f\": container with ID starting with 0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f not found: ID does not exist" containerID="0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.535119 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f"} err="failed to get container status \"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f\": rpc error: code = NotFound desc = could not find container \"0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f\": container with ID starting with 0c08d49c16fc113e90ee07ac0d1ce98c877d82c4fce4bbf69e9e4b15a851cc7f not found: ID does not exist" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.535138 4867 scope.go:117] "RemoveContainer" containerID="6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9" Dec 12 09:21:16 crc kubenswrapper[4867]: E1212 09:21:16.535583 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9\": container with ID starting with 6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9 not found: ID does not exist" containerID="6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.535615 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9"} err="failed to get container status \"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9\": rpc error: code = NotFound desc = could not find container \"6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9\": container with ID starting with 6fd2968210a9cf6394ca84007e64a1949e04bb0ebad1d9e7a370c2f286eafee9 not found: ID does not exist" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.535636 4867 scope.go:117] "RemoveContainer" containerID="589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125" Dec 12 09:21:16 crc kubenswrapper[4867]: E1212 09:21:16.536134 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125\": container with ID starting with 589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125 not found: ID does not exist" containerID="589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.536187 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125"} err="failed to get container status \"589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125\": rpc error: code = NotFound desc = could not find container \"589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125\": container with ID starting with 589fba1ad36edc26f49d64d32acb5e691de049c3f39e858623bfc3f83a327125 not found: ID does not exist" Dec 12 09:21:16 crc kubenswrapper[4867]: I1212 09:21:16.848890 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" path="/var/lib/kubelet/pods/709ec057-02f2-48ce-88ce-73c37d8d8a44/volumes" Dec 12 09:21:29 crc kubenswrapper[4867]: I1212 09:21:29.838498 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:21:30 crc kubenswrapper[4867]: I1212 09:21:30.564617 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a"} Dec 12 09:21:48 crc kubenswrapper[4867]: I1212 09:21:48.745037 4867 generic.go:334] "Generic (PLEG): container finished" podID="7307958d-09aa-4a51-b62b-7dac152de62a" containerID="1cd3002cc7cbb387954e1a1baa6fef0c0704cfadbad74bb4f351f832285e76f3" exitCode=0 Dec 12 09:21:48 crc kubenswrapper[4867]: I1212 09:21:48.745088 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" event={"ID":"7307958d-09aa-4a51-b62b-7dac152de62a","Type":"ContainerDied","Data":"1cd3002cc7cbb387954e1a1baa6fef0c0704cfadbad74bb4f351f832285e76f3"} Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.746760 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.766311 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" event={"ID":"7307958d-09aa-4a51-b62b-7dac152de62a","Type":"ContainerDied","Data":"03f5dede87006755df0443b86a8bc091144c2f1865aeee76ca68bd7388ecb0aa"} Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.766373 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03f5dede87006755df0443b86a8bc091144c2f1865aeee76ca68bd7388ecb0aa" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.766393 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.844756 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.844814 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.846565 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.846697 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.847141 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.847191 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.847251 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ml49\" (UniqueName: \"kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.847343 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.847412 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory\") pod \"7307958d-09aa-4a51-b62b-7dac152de62a\" (UID: \"7307958d-09aa-4a51-b62b-7dac152de62a\") " Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.852623 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.853357 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49" (OuterVolumeSpecName: "kube-api-access-5ml49") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "kube-api-access-5ml49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.873365 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.878292 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.880039 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.882578 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.891524 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory" (OuterVolumeSpecName: "inventory") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.897689 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.898269 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7307958d-09aa-4a51-b62b-7dac152de62a" (UID: "7307958d-09aa-4a51-b62b-7dac152de62a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952801 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952841 4867 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-inventory\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952856 4867 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952867 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952876 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952886 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952895 4867 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952906 4867 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7307958d-09aa-4a51-b62b-7dac152de62a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 12 09:21:50 crc kubenswrapper[4867]: I1212 09:21:50.952917 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ml49\" (UniqueName: \"kubernetes.io/projected/7307958d-09aa-4a51-b62b-7dac152de62a-kube-api-access-5ml49\") on node \"crc\" DevicePath \"\"" Dec 12 09:23:34 crc kubenswrapper[4867]: I1212 09:23:34.294923 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 09:23:34 crc kubenswrapper[4867]: I1212 09:23:34.295723 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="76e39b07-673e-40d8-a2ec-3f6fbd894418" containerName="adoption" containerID="cri-o://9dadee254a52c43da417caeeef0783d37846dea1fe84f9b21c4541f7826403df" gracePeriod=30 Dec 12 09:23:58 crc kubenswrapper[4867]: I1212 09:23:58.989411 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:23:58 crc kubenswrapper[4867]: I1212 09:23:58.992139 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.014202 4867 generic.go:334] "Generic (PLEG): container finished" podID="76e39b07-673e-40d8-a2ec-3f6fbd894418" containerID="9dadee254a52c43da417caeeef0783d37846dea1fe84f9b21c4541f7826403df" exitCode=137 Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.015471 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"76e39b07-673e-40d8-a2ec-3f6fbd894418","Type":"ContainerDied","Data":"9dadee254a52c43da417caeeef0783d37846dea1fe84f9b21c4541f7826403df"} Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.678855 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.748583 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") pod \"76e39b07-673e-40d8-a2ec-3f6fbd894418\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.748741 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqjrq\" (UniqueName: \"kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq\") pod \"76e39b07-673e-40d8-a2ec-3f6fbd894418\" (UID: \"76e39b07-673e-40d8-a2ec-3f6fbd894418\") " Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.759169 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq" (OuterVolumeSpecName: "kube-api-access-sqjrq") pod "76e39b07-673e-40d8-a2ec-3f6fbd894418" (UID: "76e39b07-673e-40d8-a2ec-3f6fbd894418"). InnerVolumeSpecName "kube-api-access-sqjrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.779399 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec" (OuterVolumeSpecName: "mariadb-data") pod "76e39b07-673e-40d8-a2ec-3f6fbd894418" (UID: "76e39b07-673e-40d8-a2ec-3f6fbd894418"). InnerVolumeSpecName "pvc-f48b1055-c273-46fe-b53b-59745b1711ec". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.851587 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqjrq\" (UniqueName: \"kubernetes.io/projected/76e39b07-673e-40d8-a2ec-3f6fbd894418-kube-api-access-sqjrq\") on node \"crc\" DevicePath \"\"" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.851649 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") on node \"crc\" " Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.876738 4867 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.876915 4867 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f48b1055-c273-46fe-b53b-59745b1711ec" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec") on node "crc" Dec 12 09:24:05 crc kubenswrapper[4867]: I1212 09:24:05.953883 4867 reconciler_common.go:293] "Volume detached for volume \"pvc-f48b1055-c273-46fe-b53b-59745b1711ec\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f48b1055-c273-46fe-b53b-59745b1711ec\") on node \"crc\" DevicePath \"\"" Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.026455 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"76e39b07-673e-40d8-a2ec-3f6fbd894418","Type":"ContainerDied","Data":"5026f6b8f633c9ee142eb0855d4136cfe5ee6612a218e18ec8d78a1d932e906e"} Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.026503 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.026526 4867 scope.go:117] "RemoveContainer" containerID="9dadee254a52c43da417caeeef0783d37846dea1fe84f9b21c4541f7826403df" Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.063788 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.078399 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.651771 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.652210 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="72d8033d-f19a-462d-9f0a-1122235ef5e9" containerName="adoption" containerID="cri-o://fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac" gracePeriod=30 Dec 12 09:24:06 crc kubenswrapper[4867]: I1212 09:24:06.850506 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e39b07-673e-40d8-a2ec-3f6fbd894418" path="/var/lib/kubelet/pods/76e39b07-673e-40d8-a2ec-3f6fbd894418/volumes" Dec 12 09:24:28 crc kubenswrapper[4867]: I1212 09:24:28.988573 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:24:28 crc kubenswrapper[4867]: I1212 09:24:28.989238 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.144001 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.216711 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v94gg\" (UniqueName: \"kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg\") pod \"72d8033d-f19a-462d-9f0a-1122235ef5e9\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.217152 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") pod \"72d8033d-f19a-462d-9f0a-1122235ef5e9\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.217272 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert\") pod \"72d8033d-f19a-462d-9f0a-1122235ef5e9\" (UID: \"72d8033d-f19a-462d-9f0a-1122235ef5e9\") " Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.222631 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "72d8033d-f19a-462d-9f0a-1122235ef5e9" (UID: "72d8033d-f19a-462d-9f0a-1122235ef5e9"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.231615 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg" (OuterVolumeSpecName: "kube-api-access-v94gg") pod "72d8033d-f19a-462d-9f0a-1122235ef5e9" (UID: "72d8033d-f19a-462d-9f0a-1122235ef5e9"). InnerVolumeSpecName "kube-api-access-v94gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.245884 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0" (OuterVolumeSpecName: "ovn-data") pod "72d8033d-f19a-462d-9f0a-1122235ef5e9" (UID: "72d8033d-f19a-462d-9f0a-1122235ef5e9"). InnerVolumeSpecName "pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.322100 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") on node \"crc\" " Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.322144 4867 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/72d8033d-f19a-462d-9f0a-1122235ef5e9-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.322158 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v94gg\" (UniqueName: \"kubernetes.io/projected/72d8033d-f19a-462d-9f0a-1122235ef5e9-kube-api-access-v94gg\") on node \"crc\" DevicePath \"\"" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.343552 4867 generic.go:334] "Generic (PLEG): container finished" podID="72d8033d-f19a-462d-9f0a-1122235ef5e9" containerID="fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac" exitCode=137 Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.343602 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.343603 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"72d8033d-f19a-462d-9f0a-1122235ef5e9","Type":"ContainerDied","Data":"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac"} Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.343732 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"72d8033d-f19a-462d-9f0a-1122235ef5e9","Type":"ContainerDied","Data":"d29c4b19ddad27e62789e91bd6b05256c9291659f57ca4e505201cdfb103cc2e"} Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.343757 4867 scope.go:117] "RemoveContainer" containerID="fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.350635 4867 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.350806 4867 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0") on node "crc" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.377508 4867 scope.go:117] "RemoveContainer" containerID="fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac" Dec 12 09:24:37 crc kubenswrapper[4867]: E1212 09:24:37.379658 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac\": container with ID starting with fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac not found: ID does not exist" containerID="fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.379756 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac"} err="failed to get container status \"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac\": rpc error: code = NotFound desc = could not find container \"fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac\": container with ID starting with fdf49f58df474af7476cdb8e165064647cf10c9d918c1797cf634fa3782230ac not found: ID does not exist" Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.385962 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.394520 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 12 09:24:37 crc kubenswrapper[4867]: I1212 09:24:37.423669 4867 reconciler_common.go:293] "Volume detached for volume \"pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae35aff9-842b-439c-aeb2-6b45a0d2a1f0\") on node \"crc\" DevicePath \"\"" Dec 12 09:24:38 crc kubenswrapper[4867]: I1212 09:24:38.857141 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d8033d-f19a-462d-9f0a-1122235ef5e9" path="/var/lib/kubelet/pods/72d8033d-f19a-462d-9f0a-1122235ef5e9/volumes" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.894691 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.896970 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="extract-utilities" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897072 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="extract-utilities" Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.897190 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="registry-server" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897291 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="registry-server" Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.897377 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7307958d-09aa-4a51-b62b-7dac152de62a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897452 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7307958d-09aa-4a51-b62b-7dac152de62a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.897537 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e39b07-673e-40d8-a2ec-3f6fbd894418" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897611 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e39b07-673e-40d8-a2ec-3f6fbd894418" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.897696 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="extract-content" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897767 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="extract-content" Dec 12 09:24:56 crc kubenswrapper[4867]: E1212 09:24:56.897844 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d8033d-f19a-462d-9f0a-1122235ef5e9" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.897914 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d8033d-f19a-462d-9f0a-1122235ef5e9" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.898274 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e39b07-673e-40d8-a2ec-3f6fbd894418" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.898383 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="709ec057-02f2-48ce-88ce-73c37d8d8a44" containerName="registry-server" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.898478 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7307958d-09aa-4a51-b62b-7dac152de62a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.898566 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d8033d-f19a-462d-9f0a-1122235ef5e9" containerName="adoption" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.899594 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.902032 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.902260 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.902401 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zg96v" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.903299 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 12 09:24:56 crc kubenswrapper[4867]: I1212 09:24:56.906774 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.014094 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.014487 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ljpb\" (UniqueName: \"kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.014517 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.014676 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.014988 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.015084 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.015166 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.015251 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.015303 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.118630 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.118762 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.118802 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.118868 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.119296 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.119587 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.120589 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.120714 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.120870 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ljpb\" (UniqueName: \"kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.121529 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.121827 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.123026 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.123085 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.126212 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.127479 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.132212 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.132730 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.149151 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ljpb\" (UniqueName: \"kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.164898 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"tempest-tests-tempest\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.240122 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 12 09:24:57 crc kubenswrapper[4867]: I1212 09:24:57.690148 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.541584 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ea049736-19d9-474a-8c1f-39193e6f35df","Type":"ContainerStarted","Data":"6b55d91fb191975025e27f5ea599fbb54440935ebacf1f4687dcad67e9b195bc"} Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.989451 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.989750 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.989794 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.990684 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:24:58 crc kubenswrapper[4867]: I1212 09:24:58.990754 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a" gracePeriod=600 Dec 12 09:24:59 crc kubenswrapper[4867]: I1212 09:24:59.557411 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a"} Dec 12 09:24:59 crc kubenswrapper[4867]: I1212 09:24:59.557764 4867 scope.go:117] "RemoveContainer" containerID="a0b900f7291d53eccd5fc30d8ab0ccbcb9f386feb7559dc5f3c1fee6ee29e6f6" Dec 12 09:24:59 crc kubenswrapper[4867]: I1212 09:24:59.557454 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a" exitCode=0 Dec 12 09:24:59 crc kubenswrapper[4867]: I1212 09:24:59.557878 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0"} Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.880743 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.885983 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.894206 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.933983 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.934122 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mw5m\" (UniqueName: \"kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:09 crc kubenswrapper[4867]: I1212 09:25:09.935152 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.037330 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.037427 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mw5m\" (UniqueName: \"kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.037472 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.037916 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.038218 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.059943 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mw5m\" (UniqueName: \"kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m\") pod \"redhat-marketplace-c2lkw\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.211620 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:10 crc kubenswrapper[4867]: I1212 09:25:10.859408 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:11 crc kubenswrapper[4867]: I1212 09:25:11.691762 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerStarted","Data":"06276abc2cc0b3283b1508a07308863a14729afea01a5aa85482122aefed0c95"} Dec 12 09:25:12 crc kubenswrapper[4867]: E1212 09:25:12.024507 4867 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebe2ebda_7054_489d_ae19_577525646eb8.slice/crio-dc53cb7bb80bf525151560076b0fab5c66a4d70fd195dd6d8df8dd52389a2564.scope\": RecentStats: unable to find data in memory cache]" Dec 12 09:25:12 crc kubenswrapper[4867]: I1212 09:25:12.705969 4867 generic.go:334] "Generic (PLEG): container finished" podID="ebe2ebda-7054-489d-ae19-577525646eb8" containerID="dc53cb7bb80bf525151560076b0fab5c66a4d70fd195dd6d8df8dd52389a2564" exitCode=0 Dec 12 09:25:12 crc kubenswrapper[4867]: I1212 09:25:12.706083 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerDied","Data":"dc53cb7bb80bf525151560076b0fab5c66a4d70fd195dd6d8df8dd52389a2564"} Dec 12 09:25:13 crc kubenswrapper[4867]: I1212 09:25:13.718761 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerStarted","Data":"4eb6c185143c7d296c8bfcc0117e9565920c50527a4c5a57be711493ef404d52"} Dec 12 09:25:14 crc kubenswrapper[4867]: I1212 09:25:14.733442 4867 generic.go:334] "Generic (PLEG): container finished" podID="ebe2ebda-7054-489d-ae19-577525646eb8" containerID="4eb6c185143c7d296c8bfcc0117e9565920c50527a4c5a57be711493ef404d52" exitCode=0 Dec 12 09:25:14 crc kubenswrapper[4867]: I1212 09:25:14.733542 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerDied","Data":"4eb6c185143c7d296c8bfcc0117e9565920c50527a4c5a57be711493ef404d52"} Dec 12 09:25:16 crc kubenswrapper[4867]: I1212 09:25:16.756480 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerStarted","Data":"dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe"} Dec 12 09:25:16 crc kubenswrapper[4867]: I1212 09:25:16.782351 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c2lkw" podStartSLOduration=4.100487136 podStartE2EDuration="7.782329242s" podCreationTimestamp="2025-12-12 09:25:09 +0000 UTC" firstStartedPulling="2025-12-12 09:25:12.709353662 +0000 UTC m=+9400.280734931" lastFinishedPulling="2025-12-12 09:25:16.391195768 +0000 UTC m=+9403.962577037" observedRunningTime="2025-12-12 09:25:16.771447579 +0000 UTC m=+9404.342828848" watchObservedRunningTime="2025-12-12 09:25:16.782329242 +0000 UTC m=+9404.353710511" Dec 12 09:25:20 crc kubenswrapper[4867]: I1212 09:25:20.212464 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:20 crc kubenswrapper[4867]: I1212 09:25:20.212790 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:20 crc kubenswrapper[4867]: I1212 09:25:20.270155 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:30 crc kubenswrapper[4867]: I1212 09:25:30.259006 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:30 crc kubenswrapper[4867]: I1212 09:25:30.315546 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:30 crc kubenswrapper[4867]: I1212 09:25:30.925143 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c2lkw" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="registry-server" containerID="cri-o://dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" gracePeriod=2 Dec 12 09:25:31 crc kubenswrapper[4867]: I1212 09:25:31.936366 4867 generic.go:334] "Generic (PLEG): container finished" podID="ebe2ebda-7054-489d-ae19-577525646eb8" containerID="dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" exitCode=0 Dec 12 09:25:31 crc kubenswrapper[4867]: I1212 09:25:31.936417 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerDied","Data":"dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe"} Dec 12 09:25:40 crc kubenswrapper[4867]: E1212 09:25:40.212458 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe is running failed: container process not found" containerID="dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" cmd=["grpc_health_probe","-addr=:50051"] Dec 12 09:25:40 crc kubenswrapper[4867]: E1212 09:25:40.213557 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe is running failed: container process not found" containerID="dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" cmd=["grpc_health_probe","-addr=:50051"] Dec 12 09:25:40 crc kubenswrapper[4867]: E1212 09:25:40.213974 4867 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe is running failed: container process not found" containerID="dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" cmd=["grpc_health_probe","-addr=:50051"] Dec 12 09:25:40 crc kubenswrapper[4867]: E1212 09:25:40.214013 4867 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-c2lkw" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="registry-server" Dec 12 09:25:44 crc kubenswrapper[4867]: E1212 09:25:44.236078 4867 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0a4adc83d85cf6902627a7ff9891be56" Dec 12 09:25:44 crc kubenswrapper[4867]: E1212 09:25:44.236752 4867 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0a4adc83d85cf6902627a7ff9891be56" Dec 12 09:25:44 crc kubenswrapper[4867]: E1212 09:25:44.236942 4867 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0a4adc83d85cf6902627a7ff9891be56,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ljpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ea049736-19d9-474a-8c1f-39193e6f35df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 12 09:25:44 crc kubenswrapper[4867]: E1212 09:25:44.238220 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ea049736-19d9-474a-8c1f-39193e6f35df" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.595442 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.707108 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mw5m\" (UniqueName: \"kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m\") pod \"ebe2ebda-7054-489d-ae19-577525646eb8\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.707271 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content\") pod \"ebe2ebda-7054-489d-ae19-577525646eb8\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.707323 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities\") pod \"ebe2ebda-7054-489d-ae19-577525646eb8\" (UID: \"ebe2ebda-7054-489d-ae19-577525646eb8\") " Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.707922 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities" (OuterVolumeSpecName: "utilities") pod "ebe2ebda-7054-489d-ae19-577525646eb8" (UID: "ebe2ebda-7054-489d-ae19-577525646eb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.717436 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m" (OuterVolumeSpecName: "kube-api-access-5mw5m") pod "ebe2ebda-7054-489d-ae19-577525646eb8" (UID: "ebe2ebda-7054-489d-ae19-577525646eb8"). InnerVolumeSpecName "kube-api-access-5mw5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.720486 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebe2ebda-7054-489d-ae19-577525646eb8" (UID: "ebe2ebda-7054-489d-ae19-577525646eb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.809738 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.809780 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebe2ebda-7054-489d-ae19-577525646eb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:25:44 crc kubenswrapper[4867]: I1212 09:25:44.809791 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mw5m\" (UniqueName: \"kubernetes.io/projected/ebe2ebda-7054-489d-ae19-577525646eb8-kube-api-access-5mw5m\") on node \"crc\" DevicePath \"\"" Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.059978 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c2lkw" event={"ID":"ebe2ebda-7054-489d-ae19-577525646eb8","Type":"ContainerDied","Data":"06276abc2cc0b3283b1508a07308863a14729afea01a5aa85482122aefed0c95"} Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.060401 4867 scope.go:117] "RemoveContainer" containerID="dca441b8440d8b6004f3aba6f051b7b0389f64f8fe382e3e253e096304978fbe" Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.060000 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c2lkw" Dec 12 09:25:45 crc kubenswrapper[4867]: E1212 09:25:45.062134 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0a4adc83d85cf6902627a7ff9891be56\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ea049736-19d9-474a-8c1f-39193e6f35df" Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.082403 4867 scope.go:117] "RemoveContainer" containerID="4eb6c185143c7d296c8bfcc0117e9565920c50527a4c5a57be711493ef404d52" Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.103583 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.115518 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c2lkw"] Dec 12 09:25:45 crc kubenswrapper[4867]: I1212 09:25:45.520467 4867 scope.go:117] "RemoveContainer" containerID="dc53cb7bb80bf525151560076b0fab5c66a4d70fd195dd6d8df8dd52389a2564" Dec 12 09:25:46 crc kubenswrapper[4867]: I1212 09:25:46.851770 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" path="/var/lib/kubelet/pods/ebe2ebda-7054-489d-ae19-577525646eb8/volumes" Dec 12 09:25:56 crc kubenswrapper[4867]: I1212 09:25:56.841414 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:25:57 crc kubenswrapper[4867]: I1212 09:25:57.003615 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 12 09:25:58 crc kubenswrapper[4867]: I1212 09:25:58.186570 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ea049736-19d9-474a-8c1f-39193e6f35df","Type":"ContainerStarted","Data":"8405f8c66557b39298a075c6a90339a785561809ed92359d3083ecab2e308604"} Dec 12 09:25:58 crc kubenswrapper[4867]: I1212 09:25:58.212114 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.906650858 podStartE2EDuration="1m3.212089905s" podCreationTimestamp="2025-12-12 09:24:55 +0000 UTC" firstStartedPulling="2025-12-12 09:24:57.69582163 +0000 UTC m=+9385.267202899" lastFinishedPulling="2025-12-12 09:25:57.001260677 +0000 UTC m=+9444.572641946" observedRunningTime="2025-12-12 09:25:58.206355467 +0000 UTC m=+9445.777736736" watchObservedRunningTime="2025-12-12 09:25:58.212089905 +0000 UTC m=+9445.783471174" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.393834 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:10 crc kubenswrapper[4867]: E1212 09:26:10.395798 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="registry-server" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.395908 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="registry-server" Dec 12 09:26:10 crc kubenswrapper[4867]: E1212 09:26:10.395975 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="extract-content" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.396028 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="extract-content" Dec 12 09:26:10 crc kubenswrapper[4867]: E1212 09:26:10.396086 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="extract-utilities" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.396139 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="extract-utilities" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.396426 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebe2ebda-7054-489d-ae19-577525646eb8" containerName="registry-server" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.398153 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.409659 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.446978 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.447016 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.447465 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmzmc\" (UniqueName: \"kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.557501 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.557548 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.557665 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmzmc\" (UniqueName: \"kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.558055 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.558243 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.580987 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmzmc\" (UniqueName: \"kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc\") pod \"community-operators-nt2lx\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:10 crc kubenswrapper[4867]: I1212 09:26:10.721115 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:11 crc kubenswrapper[4867]: I1212 09:26:11.158748 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:11 crc kubenswrapper[4867]: I1212 09:26:11.335506 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerStarted","Data":"bd95d7f1a01d8b307551de4a5f254f9592115b47a2b6bfe965791934a2b5cc30"} Dec 12 09:26:12 crc kubenswrapper[4867]: I1212 09:26:12.356031 4867 generic.go:334] "Generic (PLEG): container finished" podID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerID="407c61098e85f84bbcc27f1865c3f8182cb82b6c4035bc13f14a111394f6541d" exitCode=0 Dec 12 09:26:12 crc kubenswrapper[4867]: I1212 09:26:12.356360 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerDied","Data":"407c61098e85f84bbcc27f1865c3f8182cb82b6c4035bc13f14a111394f6541d"} Dec 12 09:26:13 crc kubenswrapper[4867]: I1212 09:26:13.367119 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerStarted","Data":"7fd4b779833738380d6be45606bf7f8021565af0b4852fe724e814f0f56dff68"} Dec 12 09:26:15 crc kubenswrapper[4867]: I1212 09:26:15.388184 4867 generic.go:334] "Generic (PLEG): container finished" podID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerID="7fd4b779833738380d6be45606bf7f8021565af0b4852fe724e814f0f56dff68" exitCode=0 Dec 12 09:26:15 crc kubenswrapper[4867]: I1212 09:26:15.388255 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerDied","Data":"7fd4b779833738380d6be45606bf7f8021565af0b4852fe724e814f0f56dff68"} Dec 12 09:26:16 crc kubenswrapper[4867]: I1212 09:26:16.401796 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerStarted","Data":"d2380d79ff7e69453715c66adba9276f115d81c2e105c4f7cd078d2e99aa5782"} Dec 12 09:26:16 crc kubenswrapper[4867]: I1212 09:26:16.423408 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nt2lx" podStartSLOduration=2.918027528 podStartE2EDuration="6.423385728s" podCreationTimestamp="2025-12-12 09:26:10 +0000 UTC" firstStartedPulling="2025-12-12 09:26:12.360616095 +0000 UTC m=+9459.931997364" lastFinishedPulling="2025-12-12 09:26:15.865974285 +0000 UTC m=+9463.437355564" observedRunningTime="2025-12-12 09:26:16.421762079 +0000 UTC m=+9463.993143358" watchObservedRunningTime="2025-12-12 09:26:16.423385728 +0000 UTC m=+9463.994766997" Dec 12 09:26:20 crc kubenswrapper[4867]: I1212 09:26:20.722252 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:20 crc kubenswrapper[4867]: I1212 09:26:20.722993 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:20 crc kubenswrapper[4867]: I1212 09:26:20.781497 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:21 crc kubenswrapper[4867]: I1212 09:26:21.504218 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:21 crc kubenswrapper[4867]: I1212 09:26:21.555906 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:23 crc kubenswrapper[4867]: I1212 09:26:23.477778 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nt2lx" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="registry-server" containerID="cri-o://d2380d79ff7e69453715c66adba9276f115d81c2e105c4f7cd078d2e99aa5782" gracePeriod=2 Dec 12 09:26:24 crc kubenswrapper[4867]: I1212 09:26:24.486831 4867 generic.go:334] "Generic (PLEG): container finished" podID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerID="d2380d79ff7e69453715c66adba9276f115d81c2e105c4f7cd078d2e99aa5782" exitCode=0 Dec 12 09:26:24 crc kubenswrapper[4867]: I1212 09:26:24.487211 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerDied","Data":"d2380d79ff7e69453715c66adba9276f115d81c2e105c4f7cd078d2e99aa5782"} Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.008167 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.085110 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities\") pod \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.085305 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmzmc\" (UniqueName: \"kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc\") pod \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.085382 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content\") pod \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\" (UID: \"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1\") " Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.087249 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities" (OuterVolumeSpecName: "utilities") pod "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" (UID: "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.098159 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc" (OuterVolumeSpecName: "kube-api-access-xmzmc") pod "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" (UID: "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1"). InnerVolumeSpecName "kube-api-access-xmzmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.146243 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" (UID: "6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.188275 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmzmc\" (UniqueName: \"kubernetes.io/projected/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-kube-api-access-xmzmc\") on node \"crc\" DevicePath \"\"" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.188834 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.188847 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.501424 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nt2lx" event={"ID":"6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1","Type":"ContainerDied","Data":"bd95d7f1a01d8b307551de4a5f254f9592115b47a2b6bfe965791934a2b5cc30"} Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.501490 4867 scope.go:117] "RemoveContainer" containerID="d2380d79ff7e69453715c66adba9276f115d81c2e105c4f7cd078d2e99aa5782" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.501572 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nt2lx" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.524567 4867 scope.go:117] "RemoveContainer" containerID="7fd4b779833738380d6be45606bf7f8021565af0b4852fe724e814f0f56dff68" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.551503 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.552669 4867 scope.go:117] "RemoveContainer" containerID="407c61098e85f84bbcc27f1865c3f8182cb82b6c4035bc13f14a111394f6541d" Dec 12 09:26:25 crc kubenswrapper[4867]: I1212 09:26:25.563728 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nt2lx"] Dec 12 09:26:26 crc kubenswrapper[4867]: I1212 09:26:26.854054 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" path="/var/lib/kubelet/pods/6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1/volumes" Dec 12 09:27:28 crc kubenswrapper[4867]: I1212 09:27:28.988697 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:27:28 crc kubenswrapper[4867]: I1212 09:27:28.990413 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:27:58 crc kubenswrapper[4867]: I1212 09:27:58.988578 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:27:58 crc kubenswrapper[4867]: I1212 09:27:58.989212 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:28:28 crc kubenswrapper[4867]: I1212 09:28:28.989423 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:28:28 crc kubenswrapper[4867]: I1212 09:28:28.989999 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:28:28 crc kubenswrapper[4867]: I1212 09:28:28.990046 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:28:28 crc kubenswrapper[4867]: I1212 09:28:28.990872 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:28:28 crc kubenswrapper[4867]: I1212 09:28:28.990930 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" gracePeriod=600 Dec 12 09:28:29 crc kubenswrapper[4867]: E1212 09:28:29.147942 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:28:29 crc kubenswrapper[4867]: I1212 09:28:29.630188 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" exitCode=0 Dec 12 09:28:29 crc kubenswrapper[4867]: I1212 09:28:29.630268 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0"} Dec 12 09:28:29 crc kubenswrapper[4867]: I1212 09:28:29.630320 4867 scope.go:117] "RemoveContainer" containerID="d27490db6491913b59ccc3462a5963664ec914b2e7976916ae8f9b8487d9924a" Dec 12 09:28:29 crc kubenswrapper[4867]: I1212 09:28:29.631057 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:28:29 crc kubenswrapper[4867]: E1212 09:28:29.631441 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:28:42 crc kubenswrapper[4867]: I1212 09:28:42.847055 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:28:42 crc kubenswrapper[4867]: E1212 09:28:42.848634 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:28:56 crc kubenswrapper[4867]: I1212 09:28:56.838277 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:28:56 crc kubenswrapper[4867]: E1212 09:28:56.839073 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:29:07 crc kubenswrapper[4867]: I1212 09:29:07.839546 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:29:07 crc kubenswrapper[4867]: E1212 09:29:07.840424 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:29:19 crc kubenswrapper[4867]: I1212 09:29:19.838696 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:29:19 crc kubenswrapper[4867]: E1212 09:29:19.839560 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:29:30 crc kubenswrapper[4867]: I1212 09:29:30.842559 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:29:30 crc kubenswrapper[4867]: E1212 09:29:30.843866 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:29:41 crc kubenswrapper[4867]: I1212 09:29:41.838672 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:29:41 crc kubenswrapper[4867]: E1212 09:29:41.839425 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:29:55 crc kubenswrapper[4867]: I1212 09:29:55.838145 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:29:55 crc kubenswrapper[4867]: E1212 09:29:55.839020 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.178259 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8"] Dec 12 09:30:00 crc kubenswrapper[4867]: E1212 09:30:00.179137 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="registry-server" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.179149 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="registry-server" Dec 12 09:30:00 crc kubenswrapper[4867]: E1212 09:30:00.179165 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="extract-utilities" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.179171 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="extract-utilities" Dec 12 09:30:00 crc kubenswrapper[4867]: E1212 09:30:00.179183 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="extract-content" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.179190 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="extract-content" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.179410 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0bc401-6c5f-4c1a-9a2a-584faf4bcef1" containerName="registry-server" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.180168 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.185694 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.185851 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.201824 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8"] Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.248415 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.248588 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv4dl\" (UniqueName: \"kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.248631 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.350385 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv4dl\" (UniqueName: \"kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.350460 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.350542 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.351601 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.357471 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.373113 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv4dl\" (UniqueName: \"kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl\") pod \"collect-profiles-29425530-7fdn8\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:00 crc kubenswrapper[4867]: I1212 09:30:00.512253 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:01 crc kubenswrapper[4867]: I1212 09:30:01.151035 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8"] Dec 12 09:30:01 crc kubenswrapper[4867]: I1212 09:30:01.448556 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" event={"ID":"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b","Type":"ContainerStarted","Data":"08c5b05ecf6d6fc84cb0d460893c548a215ff21a65fa7871063baa366370d401"} Dec 12 09:30:01 crc kubenswrapper[4867]: I1212 09:30:01.448842 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" event={"ID":"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b","Type":"ContainerStarted","Data":"312d790bcf45f2e30952050bc318edebd1e4061b66bccc85e951e62a62d4bcb5"} Dec 12 09:30:01 crc kubenswrapper[4867]: I1212 09:30:01.463689 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" podStartSLOduration=1.4636711230000001 podStartE2EDuration="1.463671123s" podCreationTimestamp="2025-12-12 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:30:01.460549178 +0000 UTC m=+9689.031930447" watchObservedRunningTime="2025-12-12 09:30:01.463671123 +0000 UTC m=+9689.035052392" Dec 12 09:30:02 crc kubenswrapper[4867]: I1212 09:30:02.460555 4867 generic.go:334] "Generic (PLEG): container finished" podID="05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" containerID="08c5b05ecf6d6fc84cb0d460893c548a215ff21a65fa7871063baa366370d401" exitCode=0 Dec 12 09:30:02 crc kubenswrapper[4867]: I1212 09:30:02.460720 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" event={"ID":"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b","Type":"ContainerDied","Data":"08c5b05ecf6d6fc84cb0d460893c548a215ff21a65fa7871063baa366370d401"} Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.030702 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.132638 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv4dl\" (UniqueName: \"kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl\") pod \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.132894 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume\") pod \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.133023 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume\") pod \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\" (UID: \"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b\") " Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.134268 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume" (OuterVolumeSpecName: "config-volume") pod "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" (UID: "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.141379 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" (UID: "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.141421 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl" (OuterVolumeSpecName: "kube-api-access-mv4dl") pod "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" (UID: "05bd2366-d9a4-4b2e-ae06-9be8ee21db4b"). InnerVolumeSpecName "kube-api-access-mv4dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.235408 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.235652 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.235712 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv4dl\" (UniqueName: \"kubernetes.io/projected/05bd2366-d9a4-4b2e-ae06-9be8ee21db4b-kube-api-access-mv4dl\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.522943 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" event={"ID":"05bd2366-d9a4-4b2e-ae06-9be8ee21db4b","Type":"ContainerDied","Data":"312d790bcf45f2e30952050bc318edebd1e4061b66bccc85e951e62a62d4bcb5"} Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.523217 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="312d790bcf45f2e30952050bc318edebd1e4061b66bccc85e951e62a62d4bcb5" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.523012 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425530-7fdn8" Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.562932 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch"] Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.573038 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425485-mv6ch"] Dec 12 09:30:04 crc kubenswrapper[4867]: I1212 09:30:04.852813 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd03ed01-4350-4edc-8144-0e9d4175e9e5" path="/var/lib/kubelet/pods/fd03ed01-4350-4edc-8144-0e9d4175e9e5/volumes" Dec 12 09:30:07 crc kubenswrapper[4867]: I1212 09:30:07.211460 4867 scope.go:117] "RemoveContainer" containerID="050a4e7bc0da784d8490bb4698388fe14c6de9134c272058fc4dbae19ba488c9" Dec 12 09:30:09 crc kubenswrapper[4867]: I1212 09:30:09.838271 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:30:09 crc kubenswrapper[4867]: E1212 09:30:09.838792 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.204282 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:12 crc kubenswrapper[4867]: E1212 09:30:12.205145 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" containerName="collect-profiles" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.205162 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" containerName="collect-profiles" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.205454 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="05bd2366-d9a4-4b2e-ae06-9be8ee21db4b" containerName="collect-profiles" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.207756 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.222964 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.325543 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpvns\" (UniqueName: \"kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.325650 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.325732 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.428022 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpvns\" (UniqueName: \"kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.428593 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.428885 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.429243 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.429300 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.461414 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpvns\" (UniqueName: \"kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns\") pod \"certified-operators-cqdhd\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:12 crc kubenswrapper[4867]: I1212 09:30:12.529985 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:13 crc kubenswrapper[4867]: I1212 09:30:13.159566 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:13 crc kubenswrapper[4867]: I1212 09:30:13.629499 4867 generic.go:334] "Generic (PLEG): container finished" podID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerID="0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca" exitCode=0 Dec 12 09:30:13 crc kubenswrapper[4867]: I1212 09:30:13.629594 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerDied","Data":"0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca"} Dec 12 09:30:13 crc kubenswrapper[4867]: I1212 09:30:13.629807 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerStarted","Data":"a6f5d64f3b8f09974c169db7c924698328d8c99db089c240ce080ae32727fc5e"} Dec 12 09:30:15 crc kubenswrapper[4867]: I1212 09:30:15.654407 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerStarted","Data":"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9"} Dec 12 09:30:16 crc kubenswrapper[4867]: I1212 09:30:16.670087 4867 generic.go:334] "Generic (PLEG): container finished" podID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerID="88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9" exitCode=0 Dec 12 09:30:16 crc kubenswrapper[4867]: I1212 09:30:16.670159 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerDied","Data":"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9"} Dec 12 09:30:20 crc kubenswrapper[4867]: I1212 09:30:20.708166 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerStarted","Data":"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1"} Dec 12 09:30:20 crc kubenswrapper[4867]: I1212 09:30:20.731082 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cqdhd" podStartSLOduration=3.382550527 podStartE2EDuration="8.731063245s" podCreationTimestamp="2025-12-12 09:30:12 +0000 UTC" firstStartedPulling="2025-12-12 09:30:13.632631752 +0000 UTC m=+9701.204013021" lastFinishedPulling="2025-12-12 09:30:18.98114446 +0000 UTC m=+9706.552525739" observedRunningTime="2025-12-12 09:30:20.729361684 +0000 UTC m=+9708.300742953" watchObservedRunningTime="2025-12-12 09:30:20.731063245 +0000 UTC m=+9708.302444514" Dec 12 09:30:21 crc kubenswrapper[4867]: I1212 09:30:21.839283 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:30:21 crc kubenswrapper[4867]: E1212 09:30:21.841882 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:30:22 crc kubenswrapper[4867]: I1212 09:30:22.531312 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:22 crc kubenswrapper[4867]: I1212 09:30:22.531440 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:22 crc kubenswrapper[4867]: I1212 09:30:22.593278 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:32 crc kubenswrapper[4867]: I1212 09:30:32.612980 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:32 crc kubenswrapper[4867]: I1212 09:30:32.663062 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:32 crc kubenswrapper[4867]: I1212 09:30:32.874223 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cqdhd" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="registry-server" containerID="cri-o://8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1" gracePeriod=2 Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.463122 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.561828 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpvns\" (UniqueName: \"kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns\") pod \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.561928 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities\") pod \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.562023 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content\") pod \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\" (UID: \"0f1a73fb-af85-4433-8db0-dd0fb53d7c00\") " Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.563818 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities" (OuterVolumeSpecName: "utilities") pod "0f1a73fb-af85-4433-8db0-dd0fb53d7c00" (UID: "0f1a73fb-af85-4433-8db0-dd0fb53d7c00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.569523 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns" (OuterVolumeSpecName: "kube-api-access-gpvns") pod "0f1a73fb-af85-4433-8db0-dd0fb53d7c00" (UID: "0f1a73fb-af85-4433-8db0-dd0fb53d7c00"). InnerVolumeSpecName "kube-api-access-gpvns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.616806 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f1a73fb-af85-4433-8db0-dd0fb53d7c00" (UID: "0f1a73fb-af85-4433-8db0-dd0fb53d7c00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.663585 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.663616 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpvns\" (UniqueName: \"kubernetes.io/projected/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-kube-api-access-gpvns\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.663627 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1a73fb-af85-4433-8db0-dd0fb53d7c00-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.885194 4867 generic.go:334] "Generic (PLEG): container finished" podID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerID="8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1" exitCode=0 Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.885265 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqdhd" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.885263 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerDied","Data":"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1"} Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.885330 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqdhd" event={"ID":"0f1a73fb-af85-4433-8db0-dd0fb53d7c00","Type":"ContainerDied","Data":"a6f5d64f3b8f09974c169db7c924698328d8c99db089c240ce080ae32727fc5e"} Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.885355 4867 scope.go:117] "RemoveContainer" containerID="8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.906155 4867 scope.go:117] "RemoveContainer" containerID="88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.918649 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.928874 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cqdhd"] Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.939913 4867 scope.go:117] "RemoveContainer" containerID="0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.983909 4867 scope.go:117] "RemoveContainer" containerID="8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1" Dec 12 09:30:33 crc kubenswrapper[4867]: E1212 09:30:33.984320 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1\": container with ID starting with 8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1 not found: ID does not exist" containerID="8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.984349 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1"} err="failed to get container status \"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1\": rpc error: code = NotFound desc = could not find container \"8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1\": container with ID starting with 8f0d8b595826026f16265f9af36a9dec33f99d816fabedb205cb6c65615151b1 not found: ID does not exist" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.984369 4867 scope.go:117] "RemoveContainer" containerID="88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9" Dec 12 09:30:33 crc kubenswrapper[4867]: E1212 09:30:33.984659 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9\": container with ID starting with 88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9 not found: ID does not exist" containerID="88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.984690 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9"} err="failed to get container status \"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9\": rpc error: code = NotFound desc = could not find container \"88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9\": container with ID starting with 88be4c5ee09a6d2f79bfc2d3dd14a8013d3cf92e8633c374559294a3ee9da0b9 not found: ID does not exist" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.984708 4867 scope.go:117] "RemoveContainer" containerID="0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca" Dec 12 09:30:33 crc kubenswrapper[4867]: E1212 09:30:33.985008 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca\": container with ID starting with 0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca not found: ID does not exist" containerID="0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca" Dec 12 09:30:33 crc kubenswrapper[4867]: I1212 09:30:33.985060 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca"} err="failed to get container status \"0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca\": rpc error: code = NotFound desc = could not find container \"0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca\": container with ID starting with 0be77df341f5802dca6ceb1ede828609de6dacd2d119750aac1ef5553abd77ca not found: ID does not exist" Dec 12 09:30:34 crc kubenswrapper[4867]: I1212 09:30:34.850986 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" path="/var/lib/kubelet/pods/0f1a73fb-af85-4433-8db0-dd0fb53d7c00/volumes" Dec 12 09:30:36 crc kubenswrapper[4867]: I1212 09:30:36.837657 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:30:36 crc kubenswrapper[4867]: E1212 09:30:36.837998 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:30:48 crc kubenswrapper[4867]: I1212 09:30:48.838626 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:30:48 crc kubenswrapper[4867]: E1212 09:30:48.839379 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:02 crc kubenswrapper[4867]: I1212 09:31:02.847430 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:31:02 crc kubenswrapper[4867]: E1212 09:31:02.848253 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:17 crc kubenswrapper[4867]: I1212 09:31:17.837788 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:31:17 crc kubenswrapper[4867]: E1212 09:31:17.838539 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:29 crc kubenswrapper[4867]: I1212 09:31:29.839419 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:31:29 crc kubenswrapper[4867]: E1212 09:31:29.840213 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:41 crc kubenswrapper[4867]: I1212 09:31:41.838363 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:31:41 crc kubenswrapper[4867]: E1212 09:31:41.839076 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.680547 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:31:54 crc kubenswrapper[4867]: E1212 09:31:54.681685 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="extract-content" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.681704 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="extract-content" Dec 12 09:31:54 crc kubenswrapper[4867]: E1212 09:31:54.681735 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="extract-utilities" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.681745 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="extract-utilities" Dec 12 09:31:54 crc kubenswrapper[4867]: E1212 09:31:54.681773 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="registry-server" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.681782 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="registry-server" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.682043 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1a73fb-af85-4433-8db0-dd0fb53d7c00" containerName="registry-server" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.684252 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.691741 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.793012 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.793065 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqjcz\" (UniqueName: \"kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.793119 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.838885 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:31:54 crc kubenswrapper[4867]: E1212 09:31:54.839265 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.895273 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.895334 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqjcz\" (UniqueName: \"kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.895382 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.895888 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.895994 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:54 crc kubenswrapper[4867]: I1212 09:31:54.918306 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqjcz\" (UniqueName: \"kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz\") pod \"redhat-operators-mwjqf\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:55 crc kubenswrapper[4867]: I1212 09:31:55.005238 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:31:55 crc kubenswrapper[4867]: I1212 09:31:55.563446 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:31:55 crc kubenswrapper[4867]: I1212 09:31:55.652312 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerStarted","Data":"b9850d8f27dc953c79e0b1fd204c39720953e8c394038aed6e61921ab1c106d6"} Dec 12 09:31:56 crc kubenswrapper[4867]: I1212 09:31:56.667482 4867 generic.go:334] "Generic (PLEG): container finished" podID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerID="7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795" exitCode=0 Dec 12 09:31:56 crc kubenswrapper[4867]: I1212 09:31:56.667707 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerDied","Data":"7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795"} Dec 12 09:31:56 crc kubenswrapper[4867]: I1212 09:31:56.670038 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:31:58 crc kubenswrapper[4867]: I1212 09:31:58.688159 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerStarted","Data":"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06"} Dec 12 09:32:01 crc kubenswrapper[4867]: I1212 09:32:01.714637 4867 generic.go:334] "Generic (PLEG): container finished" podID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerID="6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06" exitCode=0 Dec 12 09:32:01 crc kubenswrapper[4867]: I1212 09:32:01.714688 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerDied","Data":"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06"} Dec 12 09:32:02 crc kubenswrapper[4867]: I1212 09:32:02.725793 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerStarted","Data":"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b"} Dec 12 09:32:02 crc kubenswrapper[4867]: I1212 09:32:02.762625 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mwjqf" podStartSLOduration=3.070242817 podStartE2EDuration="8.762592466s" podCreationTimestamp="2025-12-12 09:31:54 +0000 UTC" firstStartedPulling="2025-12-12 09:31:56.669826186 +0000 UTC m=+9804.241207455" lastFinishedPulling="2025-12-12 09:32:02.362175835 +0000 UTC m=+9809.933557104" observedRunningTime="2025-12-12 09:32:02.761641372 +0000 UTC m=+9810.333022651" watchObservedRunningTime="2025-12-12 09:32:02.762592466 +0000 UTC m=+9810.333973755" Dec 12 09:32:05 crc kubenswrapper[4867]: I1212 09:32:05.006143 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:05 crc kubenswrapper[4867]: I1212 09:32:05.007145 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:06 crc kubenswrapper[4867]: I1212 09:32:06.053683 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mwjqf" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="registry-server" probeResult="failure" output=< Dec 12 09:32:06 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 09:32:06 crc kubenswrapper[4867]: > Dec 12 09:32:09 crc kubenswrapper[4867]: I1212 09:32:09.840391 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:32:09 crc kubenswrapper[4867]: E1212 09:32:09.841191 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:32:15 crc kubenswrapper[4867]: I1212 09:32:15.059063 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:15 crc kubenswrapper[4867]: I1212 09:32:15.113803 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:15 crc kubenswrapper[4867]: I1212 09:32:15.302893 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:32:16 crc kubenswrapper[4867]: I1212 09:32:16.870950 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mwjqf" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="registry-server" containerID="cri-o://768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b" gracePeriod=2 Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.524487 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.625993 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities\") pod \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.626173 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content\") pod \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.626433 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqjcz\" (UniqueName: \"kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz\") pod \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\" (UID: \"efa50409-4c3c-45f4-b10b-8ff7b3f2d681\") " Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.628837 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities" (OuterVolumeSpecName: "utilities") pod "efa50409-4c3c-45f4-b10b-8ff7b3f2d681" (UID: "efa50409-4c3c-45f4-b10b-8ff7b3f2d681"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.640657 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz" (OuterVolumeSpecName: "kube-api-access-zqjcz") pod "efa50409-4c3c-45f4-b10b-8ff7b3f2d681" (UID: "efa50409-4c3c-45f4-b10b-8ff7b3f2d681"). InnerVolumeSpecName "kube-api-access-zqjcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.728787 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqjcz\" (UniqueName: \"kubernetes.io/projected/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-kube-api-access-zqjcz\") on node \"crc\" DevicePath \"\"" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.728820 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.761355 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efa50409-4c3c-45f4-b10b-8ff7b3f2d681" (UID: "efa50409-4c3c-45f4-b10b-8ff7b3f2d681"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.830600 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efa50409-4c3c-45f4-b10b-8ff7b3f2d681-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.881534 4867 generic.go:334] "Generic (PLEG): container finished" podID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerID="768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b" exitCode=0 Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.881578 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerDied","Data":"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b"} Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.881610 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mwjqf" event={"ID":"efa50409-4c3c-45f4-b10b-8ff7b3f2d681","Type":"ContainerDied","Data":"b9850d8f27dc953c79e0b1fd204c39720953e8c394038aed6e61921ab1c106d6"} Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.881631 4867 scope.go:117] "RemoveContainer" containerID="768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.881647 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mwjqf" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.911134 4867 scope.go:117] "RemoveContainer" containerID="6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06" Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.915924 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:32:17 crc kubenswrapper[4867]: I1212 09:32:17.925921 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mwjqf"] Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.319099 4867 scope.go:117] "RemoveContainer" containerID="7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.388835 4867 scope.go:117] "RemoveContainer" containerID="768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b" Dec 12 09:32:18 crc kubenswrapper[4867]: E1212 09:32:18.389351 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b\": container with ID starting with 768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b not found: ID does not exist" containerID="768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.389382 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b"} err="failed to get container status \"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b\": rpc error: code = NotFound desc = could not find container \"768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b\": container with ID starting with 768852649c25ae4918ef8d89198b0a2107da0e842bbcce7f96d1d6d3376da46b not found: ID does not exist" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.389402 4867 scope.go:117] "RemoveContainer" containerID="6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06" Dec 12 09:32:18 crc kubenswrapper[4867]: E1212 09:32:18.389758 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06\": container with ID starting with 6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06 not found: ID does not exist" containerID="6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.389782 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06"} err="failed to get container status \"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06\": rpc error: code = NotFound desc = could not find container \"6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06\": container with ID starting with 6ede8707407518cc8aae44a3798ccc181cb6a6dd7cbc8351ea7224734d0cea06 not found: ID does not exist" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.389794 4867 scope.go:117] "RemoveContainer" containerID="7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795" Dec 12 09:32:18 crc kubenswrapper[4867]: E1212 09:32:18.390150 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795\": container with ID starting with 7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795 not found: ID does not exist" containerID="7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.390171 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795"} err="failed to get container status \"7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795\": rpc error: code = NotFound desc = could not find container \"7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795\": container with ID starting with 7596bcb3afb278dbef78d6663edc6ff3c4cb22db53ed6ef2a0034b652129f795 not found: ID does not exist" Dec 12 09:32:18 crc kubenswrapper[4867]: I1212 09:32:18.849081 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" path="/var/lib/kubelet/pods/efa50409-4c3c-45f4-b10b-8ff7b3f2d681/volumes" Dec 12 09:32:22 crc kubenswrapper[4867]: I1212 09:32:22.846024 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:32:22 crc kubenswrapper[4867]: E1212 09:32:22.846919 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:32:35 crc kubenswrapper[4867]: I1212 09:32:35.838478 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:32:35 crc kubenswrapper[4867]: E1212 09:32:35.839179 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:32:48 crc kubenswrapper[4867]: I1212 09:32:48.846308 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:32:48 crc kubenswrapper[4867]: E1212 09:32:48.847111 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:33:00 crc kubenswrapper[4867]: I1212 09:33:00.838160 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:33:00 crc kubenswrapper[4867]: E1212 09:33:00.839121 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:33:12 crc kubenswrapper[4867]: I1212 09:33:12.847682 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:33:12 crc kubenswrapper[4867]: E1212 09:33:12.848828 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:33:24 crc kubenswrapper[4867]: I1212 09:33:24.838799 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:33:24 crc kubenswrapper[4867]: E1212 09:33:24.839716 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:33:39 crc kubenswrapper[4867]: I1212 09:33:39.838314 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:33:40 crc kubenswrapper[4867]: I1212 09:33:40.640640 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574"} Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.747322 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:17 crc kubenswrapper[4867]: E1212 09:35:17.748424 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="extract-utilities" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.748438 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="extract-utilities" Dec 12 09:35:17 crc kubenswrapper[4867]: E1212 09:35:17.748462 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="registry-server" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.748469 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="registry-server" Dec 12 09:35:17 crc kubenswrapper[4867]: E1212 09:35:17.748477 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="extract-content" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.748483 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="extract-content" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.748704 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa50409-4c3c-45f4-b10b-8ff7b3f2d681" containerName="registry-server" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.752651 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.762091 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.825874 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.825929 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2qpz\" (UniqueName: \"kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.826031 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.928121 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.928338 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.928433 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2qpz\" (UniqueName: \"kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.928698 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.929520 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:17 crc kubenswrapper[4867]: I1212 09:35:17.954082 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2qpz\" (UniqueName: \"kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz\") pod \"redhat-marketplace-plzs7\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:18 crc kubenswrapper[4867]: I1212 09:35:18.082069 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:18 crc kubenswrapper[4867]: I1212 09:35:18.703978 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:19 crc kubenswrapper[4867]: I1212 09:35:19.570831 4867 generic.go:334] "Generic (PLEG): container finished" podID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerID="1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a" exitCode=0 Dec 12 09:35:19 crc kubenswrapper[4867]: I1212 09:35:19.570894 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerDied","Data":"1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a"} Dec 12 09:35:19 crc kubenswrapper[4867]: I1212 09:35:19.571178 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerStarted","Data":"ea3daa29ac9e884358a863c372f4c02be2d196dc563b2ea0695afbc6e1b45f36"} Dec 12 09:35:20 crc kubenswrapper[4867]: I1212 09:35:20.582216 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerStarted","Data":"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e"} Dec 12 09:35:21 crc kubenswrapper[4867]: I1212 09:35:21.591935 4867 generic.go:334] "Generic (PLEG): container finished" podID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerID="25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e" exitCode=0 Dec 12 09:35:21 crc kubenswrapper[4867]: I1212 09:35:21.592016 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerDied","Data":"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e"} Dec 12 09:35:22 crc kubenswrapper[4867]: I1212 09:35:22.605112 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerStarted","Data":"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903"} Dec 12 09:35:22 crc kubenswrapper[4867]: I1212 09:35:22.631141 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-plzs7" podStartSLOduration=2.874853033 podStartE2EDuration="5.63111757s" podCreationTimestamp="2025-12-12 09:35:17 +0000 UTC" firstStartedPulling="2025-12-12 09:35:19.572516679 +0000 UTC m=+10007.143897948" lastFinishedPulling="2025-12-12 09:35:22.328781216 +0000 UTC m=+10009.900162485" observedRunningTime="2025-12-12 09:35:22.626342874 +0000 UTC m=+10010.197724143" watchObservedRunningTime="2025-12-12 09:35:22.63111757 +0000 UTC m=+10010.202498839" Dec 12 09:35:28 crc kubenswrapper[4867]: I1212 09:35:28.082601 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:28 crc kubenswrapper[4867]: I1212 09:35:28.083878 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:28 crc kubenswrapper[4867]: I1212 09:35:28.130265 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:28 crc kubenswrapper[4867]: I1212 09:35:28.724667 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:28 crc kubenswrapper[4867]: I1212 09:35:28.789192 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:30 crc kubenswrapper[4867]: I1212 09:35:30.683539 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-plzs7" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="registry-server" containerID="cri-o://30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903" gracePeriod=2 Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.309378 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.380036 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content\") pod \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.380209 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2qpz\" (UniqueName: \"kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz\") pod \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.380385 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities\") pod \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\" (UID: \"e499cda2-84c7-4a58-8a3b-e2640e35e19b\") " Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.381256 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities" (OuterVolumeSpecName: "utilities") pod "e499cda2-84c7-4a58-8a3b-e2640e35e19b" (UID: "e499cda2-84c7-4a58-8a3b-e2640e35e19b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.398790 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz" (OuterVolumeSpecName: "kube-api-access-k2qpz") pod "e499cda2-84c7-4a58-8a3b-e2640e35e19b" (UID: "e499cda2-84c7-4a58-8a3b-e2640e35e19b"). InnerVolumeSpecName "kube-api-access-k2qpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.412314 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e499cda2-84c7-4a58-8a3b-e2640e35e19b" (UID: "e499cda2-84c7-4a58-8a3b-e2640e35e19b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.482571 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.482798 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e499cda2-84c7-4a58-8a3b-e2640e35e19b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.482856 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2qpz\" (UniqueName: \"kubernetes.io/projected/e499cda2-84c7-4a58-8a3b-e2640e35e19b-kube-api-access-k2qpz\") on node \"crc\" DevicePath \"\"" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.696854 4867 generic.go:334] "Generic (PLEG): container finished" podID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerID="30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903" exitCode=0 Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.696900 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerDied","Data":"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903"} Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.696935 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-plzs7" event={"ID":"e499cda2-84c7-4a58-8a3b-e2640e35e19b","Type":"ContainerDied","Data":"ea3daa29ac9e884358a863c372f4c02be2d196dc563b2ea0695afbc6e1b45f36"} Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.696955 4867 scope.go:117] "RemoveContainer" containerID="30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.696998 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-plzs7" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.748084 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.750811 4867 scope.go:117] "RemoveContainer" containerID="25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.757273 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-plzs7"] Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.785437 4867 scope.go:117] "RemoveContainer" containerID="1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.821826 4867 scope.go:117] "RemoveContainer" containerID="30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903" Dec 12 09:35:31 crc kubenswrapper[4867]: E1212 09:35:31.823260 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903\": container with ID starting with 30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903 not found: ID does not exist" containerID="30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.823294 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903"} err="failed to get container status \"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903\": rpc error: code = NotFound desc = could not find container \"30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903\": container with ID starting with 30ec7415623d8284834d5c283bc597334293140d015026dec991f3e2730d5903 not found: ID does not exist" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.823316 4867 scope.go:117] "RemoveContainer" containerID="25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e" Dec 12 09:35:31 crc kubenswrapper[4867]: E1212 09:35:31.823668 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e\": container with ID starting with 25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e not found: ID does not exist" containerID="25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.823705 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e"} err="failed to get container status \"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e\": rpc error: code = NotFound desc = could not find container \"25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e\": container with ID starting with 25a0361e018fe73acc87a92c89cc5de4fa0302a503ea619f942f853dfeca067e not found: ID does not exist" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.823731 4867 scope.go:117] "RemoveContainer" containerID="1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a" Dec 12 09:35:31 crc kubenswrapper[4867]: E1212 09:35:31.824069 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a\": container with ID starting with 1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a not found: ID does not exist" containerID="1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a" Dec 12 09:35:31 crc kubenswrapper[4867]: I1212 09:35:31.824098 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a"} err="failed to get container status \"1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a\": rpc error: code = NotFound desc = could not find container \"1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a\": container with ID starting with 1661df1e3dd6cb946d8b27cd50a9795c7d7d189c180772b9a14ca64e21ac599a not found: ID does not exist" Dec 12 09:35:32 crc kubenswrapper[4867]: I1212 09:35:32.852605 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" path="/var/lib/kubelet/pods/e499cda2-84c7-4a58-8a3b-e2640e35e19b/volumes" Dec 12 09:35:58 crc kubenswrapper[4867]: I1212 09:35:58.989323 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:35:58 crc kubenswrapper[4867]: I1212 09:35:58.989931 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.329914 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:11 crc kubenswrapper[4867]: E1212 09:36:11.330940 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="extract-content" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.330956 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="extract-content" Dec 12 09:36:11 crc kubenswrapper[4867]: E1212 09:36:11.330965 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="extract-utilities" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.330971 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="extract-utilities" Dec 12 09:36:11 crc kubenswrapper[4867]: E1212 09:36:11.331009 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="registry-server" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.331015 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="registry-server" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.331245 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e499cda2-84c7-4a58-8a3b-e2640e35e19b" containerName="registry-server" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.332915 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.349889 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.364993 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.365174 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.365255 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ctj5\" (UniqueName: \"kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.466838 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.466916 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ctj5\" (UniqueName: \"kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.466964 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.467454 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.467499 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.708586 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ctj5\" (UniqueName: \"kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5\") pod \"community-operators-pl8lx\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:11 crc kubenswrapper[4867]: I1212 09:36:11.964588 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:12 crc kubenswrapper[4867]: I1212 09:36:12.460146 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:13 crc kubenswrapper[4867]: I1212 09:36:13.087909 4867 generic.go:334] "Generic (PLEG): container finished" podID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerID="8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838" exitCode=0 Dec 12 09:36:13 crc kubenswrapper[4867]: I1212 09:36:13.088078 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerDied","Data":"8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838"} Dec 12 09:36:13 crc kubenswrapper[4867]: I1212 09:36:13.088306 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerStarted","Data":"03dffc6f8562cad38c9a98de88970764e4e2a7a6c8aeaa1f922284d86944b3b7"} Dec 12 09:36:15 crc kubenswrapper[4867]: I1212 09:36:15.113926 4867 generic.go:334] "Generic (PLEG): container finished" podID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerID="c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4" exitCode=0 Dec 12 09:36:15 crc kubenswrapper[4867]: I1212 09:36:15.114019 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerDied","Data":"c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4"} Dec 12 09:36:17 crc kubenswrapper[4867]: I1212 09:36:17.150452 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerStarted","Data":"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b"} Dec 12 09:36:17 crc kubenswrapper[4867]: I1212 09:36:17.154582 4867 generic.go:334] "Generic (PLEG): container finished" podID="ea049736-19d9-474a-8c1f-39193e6f35df" containerID="8405f8c66557b39298a075c6a90339a785561809ed92359d3083ecab2e308604" exitCode=0 Dec 12 09:36:17 crc kubenswrapper[4867]: I1212 09:36:17.154651 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ea049736-19d9-474a-8c1f-39193e6f35df","Type":"ContainerDied","Data":"8405f8c66557b39298a075c6a90339a785561809ed92359d3083ecab2e308604"} Dec 12 09:36:17 crc kubenswrapper[4867]: I1212 09:36:17.176337 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pl8lx" podStartSLOduration=2.718013514 podStartE2EDuration="6.176312299s" podCreationTimestamp="2025-12-12 09:36:11 +0000 UTC" firstStartedPulling="2025-12-12 09:36:13.092825357 +0000 UTC m=+10060.664206626" lastFinishedPulling="2025-12-12 09:36:16.551124142 +0000 UTC m=+10064.122505411" observedRunningTime="2025-12-12 09:36:17.171584064 +0000 UTC m=+10064.742965353" watchObservedRunningTime="2025-12-12 09:36:17.176312299 +0000 UTC m=+10064.747693568" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.611271 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716046 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716410 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716495 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716537 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716579 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716624 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716682 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716731 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716785 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.716827 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ljpb\" (UniqueName: \"kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb\") pod \"ea049736-19d9-474a-8c1f-39193e6f35df\" (UID: \"ea049736-19d9-474a-8c1f-39193e6f35df\") " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.717335 4867 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.717461 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data" (OuterVolumeSpecName: "config-data") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.722192 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.727565 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.731837 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb" (OuterVolumeSpecName: "kube-api-access-6ljpb") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "kube-api-access-6ljpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.745934 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.749475 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.759506 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.772857 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ea049736-19d9-474a-8c1f-39193e6f35df" (UID: "ea049736-19d9-474a-8c1f-39193e6f35df"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819011 4867 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-config-data\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819044 4867 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819053 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819063 4867 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ea049736-19d9-474a-8c1f-39193e6f35df-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819073 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ljpb\" (UniqueName: \"kubernetes.io/projected/ea049736-19d9-474a-8c1f-39193e6f35df-kube-api-access-6ljpb\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819083 4867 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ea049736-19d9-474a-8c1f-39193e6f35df-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819093 4867 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ea049736-19d9-474a-8c1f-39193e6f35df-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.819394 4867 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.842720 4867 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 12 09:36:18 crc kubenswrapper[4867]: I1212 09:36:18.921743 4867 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:19 crc kubenswrapper[4867]: I1212 09:36:19.171921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ea049736-19d9-474a-8c1f-39193e6f35df","Type":"ContainerDied","Data":"6b55d91fb191975025e27f5ea599fbb54440935ebacf1f4687dcad67e9b195bc"} Dec 12 09:36:19 crc kubenswrapper[4867]: I1212 09:36:19.171958 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b55d91fb191975025e27f5ea599fbb54440935ebacf1f4687dcad67e9b195bc" Dec 12 09:36:19 crc kubenswrapper[4867]: I1212 09:36:19.172015 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 12 09:36:21 crc kubenswrapper[4867]: I1212 09:36:21.965661 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:21 crc kubenswrapper[4867]: I1212 09:36:21.966129 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:22 crc kubenswrapper[4867]: I1212 09:36:22.025956 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:22 crc kubenswrapper[4867]: I1212 09:36:22.252241 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:22 crc kubenswrapper[4867]: I1212 09:36:22.332944 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.218073 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pl8lx" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="registry-server" containerID="cri-o://b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b" gracePeriod=2 Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.717291 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.832471 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content\") pod \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.832647 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities\") pod \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.832683 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ctj5\" (UniqueName: \"kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5\") pod \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\" (UID: \"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460\") " Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.834621 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities" (OuterVolumeSpecName: "utilities") pod "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" (UID: "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.839022 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5" (OuterVolumeSpecName: "kube-api-access-4ctj5") pod "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" (UID: "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460"). InnerVolumeSpecName "kube-api-access-4ctj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.935111 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:24 crc kubenswrapper[4867]: I1212 09:36:24.935157 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ctj5\" (UniqueName: \"kubernetes.io/projected/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-kube-api-access-4ctj5\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.065961 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" (UID: "7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.139489 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.242023 4867 generic.go:334] "Generic (PLEG): container finished" podID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerID="b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b" exitCode=0 Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.242428 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pl8lx" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.242452 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerDied","Data":"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b"} Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.245681 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pl8lx" event={"ID":"7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460","Type":"ContainerDied","Data":"03dffc6f8562cad38c9a98de88970764e4e2a7a6c8aeaa1f922284d86944b3b7"} Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.245745 4867 scope.go:117] "RemoveContainer" containerID="b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.276826 4867 scope.go:117] "RemoveContainer" containerID="c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.302814 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.315532 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pl8lx"] Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.316600 4867 scope.go:117] "RemoveContainer" containerID="8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.362018 4867 scope.go:117] "RemoveContainer" containerID="b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b" Dec 12 09:36:25 crc kubenswrapper[4867]: E1212 09:36:25.362622 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b\": container with ID starting with b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b not found: ID does not exist" containerID="b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.362666 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b"} err="failed to get container status \"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b\": rpc error: code = NotFound desc = could not find container \"b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b\": container with ID starting with b98bbb81fd753ee9a2591ed089e48f32397aa180b0fab419f7fa89274f8ac41b not found: ID does not exist" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.362692 4867 scope.go:117] "RemoveContainer" containerID="c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4" Dec 12 09:36:25 crc kubenswrapper[4867]: E1212 09:36:25.363056 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4\": container with ID starting with c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4 not found: ID does not exist" containerID="c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.363077 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4"} err="failed to get container status \"c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4\": rpc error: code = NotFound desc = could not find container \"c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4\": container with ID starting with c64786a86e62245ca0af94f4e30a06ac4a42d687b699ade808bd03faa4bbbff4 not found: ID does not exist" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.363090 4867 scope.go:117] "RemoveContainer" containerID="8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838" Dec 12 09:36:25 crc kubenswrapper[4867]: E1212 09:36:25.363476 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838\": container with ID starting with 8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838 not found: ID does not exist" containerID="8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838" Dec 12 09:36:25 crc kubenswrapper[4867]: I1212 09:36:25.363528 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838"} err="failed to get container status \"8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838\": rpc error: code = NotFound desc = could not find container \"8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838\": container with ID starting with 8b5e9aea52aad33206ae5232c172a5257815b33b473e4500c5d9c4ed97dd3838 not found: ID does not exist" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.786912 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 12 09:36:26 crc kubenswrapper[4867]: E1212 09:36:26.788467 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea049736-19d9-474a-8c1f-39193e6f35df" containerName="tempest-tests-tempest-tests-runner" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788484 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea049736-19d9-474a-8c1f-39193e6f35df" containerName="tempest-tests-tempest-tests-runner" Dec 12 09:36:26 crc kubenswrapper[4867]: E1212 09:36:26.788495 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="extract-utilities" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788502 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="extract-utilities" Dec 12 09:36:26 crc kubenswrapper[4867]: E1212 09:36:26.788520 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="extract-content" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788526 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="extract-content" Dec 12 09:36:26 crc kubenswrapper[4867]: E1212 09:36:26.788538 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="registry-server" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788543 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="registry-server" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788743 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" containerName="registry-server" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.788758 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea049736-19d9-474a-8c1f-39193e6f35df" containerName="tempest-tests-tempest-tests-runner" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.789527 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.792205 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zg96v" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.796008 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.849359 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460" path="/var/lib/kubelet/pods/7bc7f6b4-6147-48bd-8ef7-e7e7eb0b5460/volumes" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.975805 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2d92\" (UniqueName: \"kubernetes.io/projected/767fc201-7186-425f-95a3-fef4e63172ed-kube-api-access-n2d92\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:26 crc kubenswrapper[4867]: I1212 09:36:26.976045 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.077975 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2d92\" (UniqueName: \"kubernetes.io/projected/767fc201-7186-425f-95a3-fef4e63172ed-kube-api-access-n2d92\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.078204 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.078625 4867 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.094689 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2d92\" (UniqueName: \"kubernetes.io/projected/767fc201-7186-425f-95a3-fef4e63172ed-kube-api-access-n2d92\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.105283 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"767fc201-7186-425f-95a3-fef4e63172ed\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:27 crc kubenswrapper[4867]: I1212 09:36:27.119713 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 12 09:36:28 crc kubenswrapper[4867]: I1212 09:36:28.078544 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 12 09:36:28 crc kubenswrapper[4867]: W1212 09:36:28.310401 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod767fc201_7186_425f_95a3_fef4e63172ed.slice/crio-1bcd935a765b82d5c5cdf1a15fe1ac802d016e6fd5e9a8f7f9880a15a468da88 WatchSource:0}: Error finding container 1bcd935a765b82d5c5cdf1a15fe1ac802d016e6fd5e9a8f7f9880a15a468da88: Status 404 returned error can't find the container with id 1bcd935a765b82d5c5cdf1a15fe1ac802d016e6fd5e9a8f7f9880a15a468da88 Dec 12 09:36:28 crc kubenswrapper[4867]: I1212 09:36:28.988746 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:36:28 crc kubenswrapper[4867]: I1212 09:36:28.989214 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:36:29 crc kubenswrapper[4867]: I1212 09:36:29.292958 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"767fc201-7186-425f-95a3-fef4e63172ed","Type":"ContainerStarted","Data":"1bcd935a765b82d5c5cdf1a15fe1ac802d016e6fd5e9a8f7f9880a15a468da88"} Dec 12 09:36:31 crc kubenswrapper[4867]: I1212 09:36:31.311534 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"767fc201-7186-425f-95a3-fef4e63172ed","Type":"ContainerStarted","Data":"4eeb372e88daa322435810099552988ab43a671472024f7e0ccb8125475adf11"} Dec 12 09:36:31 crc kubenswrapper[4867]: I1212 09:36:31.329697 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.973057989 podStartE2EDuration="5.329676812s" podCreationTimestamp="2025-12-12 09:36:26 +0000 UTC" firstStartedPulling="2025-12-12 09:36:28.311947532 +0000 UTC m=+10075.883328801" lastFinishedPulling="2025-12-12 09:36:30.668566355 +0000 UTC m=+10078.239947624" observedRunningTime="2025-12-12 09:36:31.325153592 +0000 UTC m=+10078.896534861" watchObservedRunningTime="2025-12-12 09:36:31.329676812 +0000 UTC m=+10078.901058071" Dec 12 09:36:58 crc kubenswrapper[4867]: I1212 09:36:58.988492 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:36:58 crc kubenswrapper[4867]: I1212 09:36:58.989132 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:36:58 crc kubenswrapper[4867]: I1212 09:36:58.989187 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:36:58 crc kubenswrapper[4867]: I1212 09:36:58.990179 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:36:58 crc kubenswrapper[4867]: I1212 09:36:58.990627 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574" gracePeriod=600 Dec 12 09:36:59 crc kubenswrapper[4867]: I1212 09:36:59.592246 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574" exitCode=0 Dec 12 09:36:59 crc kubenswrapper[4867]: I1212 09:36:59.592299 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574"} Dec 12 09:36:59 crc kubenswrapper[4867]: I1212 09:36:59.592635 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a"} Dec 12 09:36:59 crc kubenswrapper[4867]: I1212 09:36:59.592664 4867 scope.go:117] "RemoveContainer" containerID="feaf716d250f87728d9c6fbe6c4c555319a22872b0a78a3175a51ca7a01a33f0" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.131093 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksppc/must-gather-r6cx8"] Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.133512 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.136717 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ksppc"/"default-dockercfg-kv82r" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.136725 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksppc"/"kube-root-ca.crt" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.137369 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ksppc"/"openshift-service-ca.crt" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.149756 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksppc/must-gather-r6cx8"] Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.255802 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxd7s\" (UniqueName: \"kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.255923 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.357539 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxd7s\" (UniqueName: \"kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.357678 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.358318 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.376496 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxd7s\" (UniqueName: \"kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s\") pod \"must-gather-r6cx8\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:36 crc kubenswrapper[4867]: I1212 09:37:36.457410 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:37:37 crc kubenswrapper[4867]: I1212 09:37:37.042331 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ksppc/must-gather-r6cx8"] Dec 12 09:37:37 crc kubenswrapper[4867]: I1212 09:37:37.058542 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:37:37 crc kubenswrapper[4867]: I1212 09:37:37.950338 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/must-gather-r6cx8" event={"ID":"1b51416a-1117-4f1c-8d12-b9fdc1c1307a","Type":"ContainerStarted","Data":"2dd85a55c6bc075c066de9ae3cd7783113ca49ecf0b10358330d1ab1e0c02ce5"} Dec 12 09:37:45 crc kubenswrapper[4867]: I1212 09:37:45.028997 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/must-gather-r6cx8" event={"ID":"1b51416a-1117-4f1c-8d12-b9fdc1c1307a","Type":"ContainerStarted","Data":"a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da"} Dec 12 09:37:45 crc kubenswrapper[4867]: I1212 09:37:45.029588 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/must-gather-r6cx8" event={"ID":"1b51416a-1117-4f1c-8d12-b9fdc1c1307a","Type":"ContainerStarted","Data":"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2"} Dec 12 09:37:45 crc kubenswrapper[4867]: I1212 09:37:45.046171 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ksppc/must-gather-r6cx8" podStartSLOduration=1.5909459799999999 podStartE2EDuration="9.046147031s" podCreationTimestamp="2025-12-12 09:37:36 +0000 UTC" firstStartedPulling="2025-12-12 09:37:37.058217433 +0000 UTC m=+10144.629598702" lastFinishedPulling="2025-12-12 09:37:44.513418474 +0000 UTC m=+10152.084799753" observedRunningTime="2025-12-12 09:37:45.043621059 +0000 UTC m=+10152.615002338" watchObservedRunningTime="2025-12-12 09:37:45.046147031 +0000 UTC m=+10152.617528300" Dec 12 09:37:48 crc kubenswrapper[4867]: I1212 09:37:48.986926 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksppc/crc-debug-hv8lj"] Dec 12 09:37:48 crc kubenswrapper[4867]: I1212 09:37:48.988651 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.156498 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q9hf\" (UniqueName: \"kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.157064 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.258512 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q9hf\" (UniqueName: \"kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.258556 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.258962 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.284830 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q9hf\" (UniqueName: \"kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf\") pod \"crc-debug-hv8lj\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: I1212 09:37:49.308740 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:37:49 crc kubenswrapper[4867]: W1212 09:37:49.372996 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48aabf68_4b34_4c79_a2a3_0382276af941.slice/crio-b9f3aa666d62d7c57eaed96349a22ce09c49ae3dd5ffa027789e722ff57ea8db WatchSource:0}: Error finding container b9f3aa666d62d7c57eaed96349a22ce09c49ae3dd5ffa027789e722ff57ea8db: Status 404 returned error can't find the container with id b9f3aa666d62d7c57eaed96349a22ce09c49ae3dd5ffa027789e722ff57ea8db Dec 12 09:37:50 crc kubenswrapper[4867]: I1212 09:37:50.072515 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" event={"ID":"48aabf68-4b34-4c79-a2a3-0382276af941","Type":"ContainerStarted","Data":"b9f3aa666d62d7c57eaed96349a22ce09c49ae3dd5ffa027789e722ff57ea8db"} Dec 12 09:38:04 crc kubenswrapper[4867]: I1212 09:38:04.215552 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" event={"ID":"48aabf68-4b34-4c79-a2a3-0382276af941","Type":"ContainerStarted","Data":"43ef36457cde94731bebf7f8d819341eccbb31dc0b062eb533ebb93129ebafa6"} Dec 12 09:38:59 crc kubenswrapper[4867]: I1212 09:38:59.779644 4867 generic.go:334] "Generic (PLEG): container finished" podID="48aabf68-4b34-4c79-a2a3-0382276af941" containerID="43ef36457cde94731bebf7f8d819341eccbb31dc0b062eb533ebb93129ebafa6" exitCode=0 Dec 12 09:38:59 crc kubenswrapper[4867]: I1212 09:38:59.779729 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" event={"ID":"48aabf68-4b34-4c79-a2a3-0382276af941","Type":"ContainerDied","Data":"43ef36457cde94731bebf7f8d819341eccbb31dc0b062eb533ebb93129ebafa6"} Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.898274 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.932831 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-hv8lj"] Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.941746 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-hv8lj"] Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.990578 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host\") pod \"48aabf68-4b34-4c79-a2a3-0382276af941\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.990679 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host" (OuterVolumeSpecName: "host") pod "48aabf68-4b34-4c79-a2a3-0382276af941" (UID: "48aabf68-4b34-4c79-a2a3-0382276af941"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.990937 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q9hf\" (UniqueName: \"kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf\") pod \"48aabf68-4b34-4c79-a2a3-0382276af941\" (UID: \"48aabf68-4b34-4c79-a2a3-0382276af941\") " Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.991525 4867 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48aabf68-4b34-4c79-a2a3-0382276af941-host\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:00 crc kubenswrapper[4867]: I1212 09:39:00.996533 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf" (OuterVolumeSpecName: "kube-api-access-8q9hf") pod "48aabf68-4b34-4c79-a2a3-0382276af941" (UID: "48aabf68-4b34-4c79-a2a3-0382276af941"). InnerVolumeSpecName "kube-api-access-8q9hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:39:01 crc kubenswrapper[4867]: I1212 09:39:01.093349 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q9hf\" (UniqueName: \"kubernetes.io/projected/48aabf68-4b34-4c79-a2a3-0382276af941-kube-api-access-8q9hf\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:01 crc kubenswrapper[4867]: I1212 09:39:01.803586 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9f3aa666d62d7c57eaed96349a22ce09c49ae3dd5ffa027789e722ff57ea8db" Dec 12 09:39:01 crc kubenswrapper[4867]: I1212 09:39:01.803691 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-hv8lj" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.106391 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksppc/crc-debug-mdqj7"] Dec 12 09:39:02 crc kubenswrapper[4867]: E1212 09:39:02.107090 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48aabf68-4b34-4c79-a2a3-0382276af941" containerName="container-00" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.107102 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="48aabf68-4b34-4c79-a2a3-0382276af941" containerName="container-00" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.107322 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="48aabf68-4b34-4c79-a2a3-0382276af941" containerName="container-00" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.107934 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.216101 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.216338 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lskv\" (UniqueName: \"kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.317847 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lskv\" (UniqueName: \"kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.317982 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.318115 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.334708 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lskv\" (UniqueName: \"kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv\") pod \"crc-debug-mdqj7\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.426962 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.815980 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" event={"ID":"bd09348c-018b-4331-8cc7-391e572f75e4","Type":"ContainerStarted","Data":"ac603494517619aefe6e568808eb21939b2e39e233087b6bf515146685bf60f5"} Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.816645 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" event={"ID":"bd09348c-018b-4331-8cc7-391e572f75e4","Type":"ContainerStarted","Data":"782e369be7ca97ac7c9e8f43f17ef334419fefc819b7e820398a8fc6bb9ea851"} Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.836644 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" podStartSLOduration=0.836621044 podStartE2EDuration="836.621044ms" podCreationTimestamp="2025-12-12 09:39:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:39:02.831846466 +0000 UTC m=+10230.403227745" watchObservedRunningTime="2025-12-12 09:39:02.836621044 +0000 UTC m=+10230.408002313" Dec 12 09:39:02 crc kubenswrapper[4867]: I1212 09:39:02.869089 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48aabf68-4b34-4c79-a2a3-0382276af941" path="/var/lib/kubelet/pods/48aabf68-4b34-4c79-a2a3-0382276af941/volumes" Dec 12 09:39:03 crc kubenswrapper[4867]: I1212 09:39:03.827620 4867 generic.go:334] "Generic (PLEG): container finished" podID="bd09348c-018b-4331-8cc7-391e572f75e4" containerID="ac603494517619aefe6e568808eb21939b2e39e233087b6bf515146685bf60f5" exitCode=0 Dec 12 09:39:03 crc kubenswrapper[4867]: I1212 09:39:03.827950 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" event={"ID":"bd09348c-018b-4331-8cc7-391e572f75e4","Type":"ContainerDied","Data":"ac603494517619aefe6e568808eb21939b2e39e233087b6bf515146685bf60f5"} Dec 12 09:39:04 crc kubenswrapper[4867]: I1212 09:39:04.953463 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.065331 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host\") pod \"bd09348c-018b-4331-8cc7-391e572f75e4\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.065371 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lskv\" (UniqueName: \"kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv\") pod \"bd09348c-018b-4331-8cc7-391e572f75e4\" (UID: \"bd09348c-018b-4331-8cc7-391e572f75e4\") " Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.065563 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host" (OuterVolumeSpecName: "host") pod "bd09348c-018b-4331-8cc7-391e572f75e4" (UID: "bd09348c-018b-4331-8cc7-391e572f75e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.066122 4867 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd09348c-018b-4331-8cc7-391e572f75e4-host\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.072579 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv" (OuterVolumeSpecName: "kube-api-access-8lskv") pod "bd09348c-018b-4331-8cc7-391e572f75e4" (UID: "bd09348c-018b-4331-8cc7-391e572f75e4"). InnerVolumeSpecName "kube-api-access-8lskv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.167593 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lskv\" (UniqueName: \"kubernetes.io/projected/bd09348c-018b-4331-8cc7-391e572f75e4-kube-api-access-8lskv\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.420583 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-mdqj7"] Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.435579 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-mdqj7"] Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.845730 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="782e369be7ca97ac7c9e8f43f17ef334419fefc819b7e820398a8fc6bb9ea851" Dec 12 09:39:05 crc kubenswrapper[4867]: I1212 09:39:05.846101 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-mdqj7" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.599001 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ksppc/crc-debug-qsghb"] Dec 12 09:39:06 crc kubenswrapper[4867]: E1212 09:39:06.600108 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd09348c-018b-4331-8cc7-391e572f75e4" containerName="container-00" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.600127 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd09348c-018b-4331-8cc7-391e572f75e4" containerName="container-00" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.600518 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd09348c-018b-4331-8cc7-391e572f75e4" containerName="container-00" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.601529 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.724958 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6lpz\" (UniqueName: \"kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.725418 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.827825 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.827998 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.828056 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6lpz\" (UniqueName: \"kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.848871 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd09348c-018b-4331-8cc7-391e572f75e4" path="/var/lib/kubelet/pods/bd09348c-018b-4331-8cc7-391e572f75e4/volumes" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.850157 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6lpz\" (UniqueName: \"kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz\") pod \"crc-debug-qsghb\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: I1212 09:39:06.920420 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:06 crc kubenswrapper[4867]: W1212 09:39:06.947484 4867 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda36aff8_917e_4d56_896f_a5aaad0953b2.slice/crio-690320084c3fdf40afcb20f3f2cf203804eb875153afe5e46c9322dd6547cd44 WatchSource:0}: Error finding container 690320084c3fdf40afcb20f3f2cf203804eb875153afe5e46c9322dd6547cd44: Status 404 returned error can't find the container with id 690320084c3fdf40afcb20f3f2cf203804eb875153afe5e46c9322dd6547cd44 Dec 12 09:39:07 crc kubenswrapper[4867]: I1212 09:39:07.863241 4867 generic.go:334] "Generic (PLEG): container finished" podID="da36aff8-917e-4d56-896f-a5aaad0953b2" containerID="b9bb7c569becf0c673e8e0cca1facdc905ee724c295e77aecbf143f3c75dd873" exitCode=0 Dec 12 09:39:07 crc kubenswrapper[4867]: I1212 09:39:07.863265 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-qsghb" event={"ID":"da36aff8-917e-4d56-896f-a5aaad0953b2","Type":"ContainerDied","Data":"b9bb7c569becf0c673e8e0cca1facdc905ee724c295e77aecbf143f3c75dd873"} Dec 12 09:39:07 crc kubenswrapper[4867]: I1212 09:39:07.863671 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/crc-debug-qsghb" event={"ID":"da36aff8-917e-4d56-896f-a5aaad0953b2","Type":"ContainerStarted","Data":"690320084c3fdf40afcb20f3f2cf203804eb875153afe5e46c9322dd6547cd44"} Dec 12 09:39:07 crc kubenswrapper[4867]: I1212 09:39:07.905500 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-qsghb"] Dec 12 09:39:07 crc kubenswrapper[4867]: I1212 09:39:07.915364 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksppc/crc-debug-qsghb"] Dec 12 09:39:08 crc kubenswrapper[4867]: I1212 09:39:08.972655 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.067457 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6lpz\" (UniqueName: \"kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz\") pod \"da36aff8-917e-4d56-896f-a5aaad0953b2\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.067742 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host\") pod \"da36aff8-917e-4d56-896f-a5aaad0953b2\" (UID: \"da36aff8-917e-4d56-896f-a5aaad0953b2\") " Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.068490 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host" (OuterVolumeSpecName: "host") pod "da36aff8-917e-4d56-896f-a5aaad0953b2" (UID: "da36aff8-917e-4d56-896f-a5aaad0953b2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.068621 4867 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da36aff8-917e-4d56-896f-a5aaad0953b2-host\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.074179 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz" (OuterVolumeSpecName: "kube-api-access-t6lpz") pod "da36aff8-917e-4d56-896f-a5aaad0953b2" (UID: "da36aff8-917e-4d56-896f-a5aaad0953b2"). InnerVolumeSpecName "kube-api-access-t6lpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.170509 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6lpz\" (UniqueName: \"kubernetes.io/projected/da36aff8-917e-4d56-896f-a5aaad0953b2-kube-api-access-t6lpz\") on node \"crc\" DevicePath \"\"" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.880844 4867 scope.go:117] "RemoveContainer" containerID="b9bb7c569becf0c673e8e0cca1facdc905ee724c295e77aecbf143f3c75dd873" Dec 12 09:39:09 crc kubenswrapper[4867]: I1212 09:39:09.880899 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/crc-debug-qsghb" Dec 12 09:39:10 crc kubenswrapper[4867]: I1212 09:39:10.852041 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da36aff8-917e-4d56-896f-a5aaad0953b2" path="/var/lib/kubelet/pods/da36aff8-917e-4d56-896f-a5aaad0953b2/volumes" Dec 12 09:39:28 crc kubenswrapper[4867]: I1212 09:39:28.990344 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:39:28 crc kubenswrapper[4867]: I1212 09:39:28.991016 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:39:58 crc kubenswrapper[4867]: I1212 09:39:58.988766 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:39:58 crc kubenswrapper[4867]: I1212 09:39:58.989268 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:40:28 crc kubenswrapper[4867]: I1212 09:40:28.988853 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:40:28 crc kubenswrapper[4867]: I1212 09:40:28.989710 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:40:28 crc kubenswrapper[4867]: I1212 09:40:28.989763 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:40:28 crc kubenswrapper[4867]: I1212 09:40:28.990803 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:40:28 crc kubenswrapper[4867]: I1212 09:40:28.990868 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" gracePeriod=600 Dec 12 09:40:29 crc kubenswrapper[4867]: E1212 09:40:29.118055 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:40:29 crc kubenswrapper[4867]: I1212 09:40:29.611650 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" exitCode=0 Dec 12 09:40:29 crc kubenswrapper[4867]: I1212 09:40:29.611708 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a"} Dec 12 09:40:29 crc kubenswrapper[4867]: I1212 09:40:29.611743 4867 scope.go:117] "RemoveContainer" containerID="d52e6a00653d9fd80eb279271cfe3b992f016f8c14488c9825db5392fedde574" Dec 12 09:40:29 crc kubenswrapper[4867]: I1212 09:40:29.612510 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:40:29 crc kubenswrapper[4867]: E1212 09:40:29.612773 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.827514 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:33 crc kubenswrapper[4867]: E1212 09:40:33.828724 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da36aff8-917e-4d56-896f-a5aaad0953b2" containerName="container-00" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.828742 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="da36aff8-917e-4d56-896f-a5aaad0953b2" containerName="container-00" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.828991 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="da36aff8-917e-4d56-896f-a5aaad0953b2" containerName="container-00" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.830980 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.842692 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.929326 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.929417 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:33 crc kubenswrapper[4867]: I1212 09:40:33.929567 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6rnb\" (UniqueName: \"kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.031720 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6rnb\" (UniqueName: \"kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.031886 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.031949 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.032545 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.032588 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.054290 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6rnb\" (UniqueName: \"kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb\") pod \"certified-operators-fsfss\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.198778 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:34 crc kubenswrapper[4867]: I1212 09:40:34.764222 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:35 crc kubenswrapper[4867]: I1212 09:40:35.670940 4867 generic.go:334] "Generic (PLEG): container finished" podID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerID="08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449" exitCode=0 Dec 12 09:40:35 crc kubenswrapper[4867]: I1212 09:40:35.671203 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerDied","Data":"08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449"} Dec 12 09:40:35 crc kubenswrapper[4867]: I1212 09:40:35.671246 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerStarted","Data":"14073f462a45fcb869f379ef390897248c01108616d8c17e4af155b452a052b1"} Dec 12 09:40:36 crc kubenswrapper[4867]: I1212 09:40:36.682484 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerStarted","Data":"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87"} Dec 12 09:40:37 crc kubenswrapper[4867]: I1212 09:40:37.697888 4867 generic.go:334] "Generic (PLEG): container finished" podID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerID="a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87" exitCode=0 Dec 12 09:40:37 crc kubenswrapper[4867]: I1212 09:40:37.698859 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerDied","Data":"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87"} Dec 12 09:40:38 crc kubenswrapper[4867]: I1212 09:40:38.711126 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerStarted","Data":"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892"} Dec 12 09:40:38 crc kubenswrapper[4867]: I1212 09:40:38.730083 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fsfss" podStartSLOduration=3.194052641 podStartE2EDuration="5.730064215s" podCreationTimestamp="2025-12-12 09:40:33 +0000 UTC" firstStartedPulling="2025-12-12 09:40:35.673167057 +0000 UTC m=+10323.244548326" lastFinishedPulling="2025-12-12 09:40:38.209178641 +0000 UTC m=+10325.780559900" observedRunningTime="2025-12-12 09:40:38.729172452 +0000 UTC m=+10326.300553731" watchObservedRunningTime="2025-12-12 09:40:38.730064215 +0000 UTC m=+10326.301445484" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.199008 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.199593 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.252727 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.804093 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.839307 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:40:44 crc kubenswrapper[4867]: E1212 09:40:44.840264 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:40:44 crc kubenswrapper[4867]: I1212 09:40:44.849866 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:46 crc kubenswrapper[4867]: I1212 09:40:46.779887 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fsfss" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="registry-server" containerID="cri-o://50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892" gracePeriod=2 Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.233264 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.299052 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities\") pod \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.299515 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6rnb\" (UniqueName: \"kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb\") pod \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.299657 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content\") pod \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\" (UID: \"e931a3c8-10b2-4cb9-8145-e9111a6da0a6\") " Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.300547 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities" (OuterVolumeSpecName: "utilities") pod "e931a3c8-10b2-4cb9-8145-e9111a6da0a6" (UID: "e931a3c8-10b2-4cb9-8145-e9111a6da0a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.309698 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb" (OuterVolumeSpecName: "kube-api-access-c6rnb") pod "e931a3c8-10b2-4cb9-8145-e9111a6da0a6" (UID: "e931a3c8-10b2-4cb9-8145-e9111a6da0a6"). InnerVolumeSpecName "kube-api-access-c6rnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.401305 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6rnb\" (UniqueName: \"kubernetes.io/projected/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-kube-api-access-c6rnb\") on node \"crc\" DevicePath \"\"" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.401338 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.525585 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e931a3c8-10b2-4cb9-8145-e9111a6da0a6" (UID: "e931a3c8-10b2-4cb9-8145-e9111a6da0a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.605490 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e931a3c8-10b2-4cb9-8145-e9111a6da0a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.793187 4867 generic.go:334] "Generic (PLEG): container finished" podID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerID="50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892" exitCode=0 Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.793256 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerDied","Data":"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892"} Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.793283 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fsfss" event={"ID":"e931a3c8-10b2-4cb9-8145-e9111a6da0a6","Type":"ContainerDied","Data":"14073f462a45fcb869f379ef390897248c01108616d8c17e4af155b452a052b1"} Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.793302 4867 scope.go:117] "RemoveContainer" containerID="50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.793322 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fsfss" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.821863 4867 scope.go:117] "RemoveContainer" containerID="a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.848758 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.849970 4867 scope.go:117] "RemoveContainer" containerID="08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.858895 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fsfss"] Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.894733 4867 scope.go:117] "RemoveContainer" containerID="50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892" Dec 12 09:40:47 crc kubenswrapper[4867]: E1212 09:40:47.895344 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892\": container with ID starting with 50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892 not found: ID does not exist" containerID="50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.895399 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892"} err="failed to get container status \"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892\": rpc error: code = NotFound desc = could not find container \"50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892\": container with ID starting with 50975836f3548390b93073bb503f8e10b56960dba5d8073bcba446fb04ae9892 not found: ID does not exist" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.895428 4867 scope.go:117] "RemoveContainer" containerID="a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87" Dec 12 09:40:47 crc kubenswrapper[4867]: E1212 09:40:47.895720 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87\": container with ID starting with a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87 not found: ID does not exist" containerID="a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.895757 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87"} err="failed to get container status \"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87\": rpc error: code = NotFound desc = could not find container \"a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87\": container with ID starting with a21d53c3d25c1e38320456f6fe15a5ffa52fb126053a45749a99ca9b71f67d87 not found: ID does not exist" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.895783 4867 scope.go:117] "RemoveContainer" containerID="08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449" Dec 12 09:40:47 crc kubenswrapper[4867]: E1212 09:40:47.896056 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449\": container with ID starting with 08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449 not found: ID does not exist" containerID="08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449" Dec 12 09:40:47 crc kubenswrapper[4867]: I1212 09:40:47.896080 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449"} err="failed to get container status \"08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449\": rpc error: code = NotFound desc = could not find container \"08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449\": container with ID starting with 08a2840da6294d219dd9d41ce63443a00d028e25c09b53ec89c730c5425e7449 not found: ID does not exist" Dec 12 09:40:48 crc kubenswrapper[4867]: I1212 09:40:48.851718 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" path="/var/lib/kubelet/pods/e931a3c8-10b2-4cb9-8145-e9111a6da0a6/volumes" Dec 12 09:40:54 crc kubenswrapper[4867]: I1212 09:40:54.635477 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b95f4529-9df6-48cf-a611-4b0b049ca8f4/init-config-reloader/0.log" Dec 12 09:40:54 crc kubenswrapper[4867]: I1212 09:40:54.804850 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b95f4529-9df6-48cf-a611-4b0b049ca8f4/init-config-reloader/0.log" Dec 12 09:40:54 crc kubenswrapper[4867]: I1212 09:40:54.850289 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b95f4529-9df6-48cf-a611-4b0b049ca8f4/alertmanager/0.log" Dec 12 09:40:54 crc kubenswrapper[4867]: I1212 09:40:54.897748 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_b95f4529-9df6-48cf-a611-4b0b049ca8f4/config-reloader/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.057610 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0b99c63-9b5a-4ffe-92f4-5fef96b3df72/aodh-api/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.103891 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0b99c63-9b5a-4ffe-92f4-5fef96b3df72/aodh-evaluator/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.263342 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0b99c63-9b5a-4ffe-92f4-5fef96b3df72/aodh-listener/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.273413 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d0b99c63-9b5a-4ffe-92f4-5fef96b3df72/aodh-notifier/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.341495 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cbd859876-xpw6c_7a54eb8d-8507-4f32-b2bf-936ea487c701/barbican-api/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.476681 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-cbd859876-xpw6c_7a54eb8d-8507-4f32-b2bf-936ea487c701/barbican-api-log/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.587525 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5bff489498-h54sr_e41196ac-a7c9-47fa-9ca0-584a5fad129b/barbican-keystone-listener/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.830484 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84b67f6bf7-jqpkj_2ad1843a-0c00-40d2-95b5-4f164cc54053/barbican-worker-log/0.log" Dec 12 09:40:55 crc kubenswrapper[4867]: I1212 09:40:55.869275 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-84b67f6bf7-jqpkj_2ad1843a-0c00-40d2-95b5-4f164cc54053/barbican-worker/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.007631 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5bff489498-h54sr_e41196ac-a7c9-47fa-9ca0-584a5fad129b/barbican-keystone-listener-log/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.117576 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-q2znc_53b9ceda-53ce-4426-9075-83f190cc8351/bootstrap-openstack-openstack-cell1/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.279430 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_50dd7d57-3939-45db-bb38-9e6198ba59be/ceilometer-central-agent/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.363096 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_50dd7d57-3939-45db-bb38-9e6198ba59be/ceilometer-notification-agent/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.377804 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_50dd7d57-3939-45db-bb38-9e6198ba59be/sg-core/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.388849 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_50dd7d57-3939-45db-bb38-9e6198ba59be/proxy-httpd/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.593768 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a7368375-0f30-4a7f-9905-7786ed91be0b/cinder-api-log/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.682420 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_a7368375-0f30-4a7f-9905-7786ed91be0b/cinder-api/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.846885 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bf535deb-d3aa-48ae-9c8d-190d12647967/probe/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.890987 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_bf535deb-d3aa-48ae-9c8d-190d12647967/cinder-scheduler/0.log" Dec 12 09:40:56 crc kubenswrapper[4867]: I1212 09:40:56.930710 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-vnwc7_1220a9e0-08bf-4499-b17a-0e9d53ca3a66/configure-network-openstack-openstack-cell1/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.107751 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-s6wxg_88b38f78-a939-4410-8b4b-a584d7a12f34/configure-os-openstack-openstack-cell1/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.130573 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69b4d64989-z2k6r_671819cf-f023-49bf-9e27-0a3e0c1ec578/init/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.407005 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69b4d64989-z2k6r_671819cf-f023-49bf-9e27-0a3e0c1ec578/dnsmasq-dns/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.436978 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-69b4d64989-z2k6r_671819cf-f023-49bf-9e27-0a3e0c1ec578/init/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.530377 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-lfg7r_146d44cb-ac9b-481c-8aa5-0884c7025d47/download-cache-openstack-openstack-cell1/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.708840 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aea8bd9c-4959-457f-b87e-dff93c91f4e7/glance-httpd/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.832428 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_aea8bd9c-4959-457f-b87e-dff93c91f4e7/glance-log/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.905593 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7630d135-1664-41e2-9b9e-76a325f47863/glance-httpd/0.log" Dec 12 09:40:57 crc kubenswrapper[4867]: I1212 09:40:57.971381 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7630d135-1664-41e2-9b9e-76a325f47863/glance-log/0.log" Dec 12 09:40:58 crc kubenswrapper[4867]: I1212 09:40:58.504739 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6dfc547d9b-s2cdv_642472d4-e097-4004-bbd1-cbb5f11fc76d/heat-engine/0.log" Dec 12 09:40:58 crc kubenswrapper[4867]: I1212 09:40:58.723471 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7b64b9cdf4-96grl_1e214fa4-0f89-41fa-8e9b-d064adcd4c3c/heat-api/0.log" Dec 12 09:40:58 crc kubenswrapper[4867]: I1212 09:40:58.868930 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f8df8656d-5rfkc_826ddec1-f8e5-41fc-bf66-a58f07371c55/horizon/0.log" Dec 12 09:40:58 crc kubenswrapper[4867]: I1212 09:40:58.979911 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-jxrx8_3ec8f159-fc91-4347-be68-266e1f8b18e1/install-certs-openstack-openstack-cell1/0.log" Dec 12 09:40:58 crc kubenswrapper[4867]: I1212 09:40:58.987955 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-864798f469-5rmb7_975458e4-1347-4cd8-8391-a7a95dfda548/heat-cfnapi/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.354755 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-99v58_a37d16ec-f91d-410d-a88a-ba1c9fc3fdc8/install-os-openstack-openstack-cell1/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.576499 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29425501-xj6cb_670929f8-dbbc-402a-94f1-6d8def78eace/keystone-cron/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.589902 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f8df8656d-5rfkc_826ddec1-f8e5-41fc-bf66-a58f07371c55/horizon-log/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.790206 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_94a7fe15-90d4-483c-b893-5c69e8a969da/kube-state-metrics/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.834297 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-x52kw_3cdd20a7-5b34-4ab7-a93a-8b8f273535e7/libvirt-openstack-openstack-cell1/0.log" Dec 12 09:40:59 crc kubenswrapper[4867]: I1212 09:40:59.839971 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:40:59 crc kubenswrapper[4867]: E1212 09:40:59.840196 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.009668 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5789865856-hkfwx_864c3801-1b19-46fb-8564-8e2d772963bc/keystone-api/0.log" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.696465 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5858cbcf-825r6_f04c345a-28e4-412e-8d33-82d2f59586db/neutron-api/0.log" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.706833 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6c5858cbcf-825r6_f04c345a-28e4-412e-8d33-82d2f59586db/neutron-httpd/0.log" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.748113 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-87rk6_61a43cf7-2462-4a34-a6f8-4c95a7571df5/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.864568 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_01e9eb2a-e9ab-4fc6-980f-0cd5367e9c6a/memcached/0.log" Dec 12 09:41:00 crc kubenswrapper[4867]: I1212 09:41:00.971248 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-h4zwq_9a40e922-5701-4ecf-b258-c8c129fd6ca1/neutron-metadata-openstack-openstack-cell1/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.054635 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-5r8zv_8dce4a04-d8b6-4c8b-b180-a360add31127/neutron-sriov-openstack-openstack-cell1/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.527136 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4065a099-8290-4414-bbc5-58c3b4ccae03/nova-cell0-conductor-conductor/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.529039 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_31e46da1-e8b1-4279-8410-9c32bea377de/nova-api-log/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.705419 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5c642a3e-a465-4936-8da1-55a2db56509b/nova-cell1-conductor-conductor/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.742105 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_31e46da1-e8b1-4279-8410-9c32bea377de/nova-api-api/0.log" Dec 12 09:41:01 crc kubenswrapper[4867]: I1212 09:41:01.840470 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4700bc38-07a1-41ea-984c-a51b6db55929/nova-cell1-novncproxy-novncproxy/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.026255 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellj9f9d_7307958d-09aa-4a51-b62b-7dac152de62a/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.190087 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-qtdct_7840c155-9e2e-4893-b6df-f81ed09aaaeb/nova-cell1-openstack-openstack-cell1/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.283563 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_97b35086-f8bc-4035-9685-212989ef3b77/nova-metadata-log/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.490987 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2c501b05-328a-4dcc-b688-ff4b2c1c90c3/nova-scheduler-scheduler/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.583307 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_07166244-e179-4d78-a254-942fcf7bee79/mysql-bootstrap/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.757815 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_07166244-e179-4d78-a254-942fcf7bee79/mysql-bootstrap/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.763514 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_07166244-e179-4d78-a254-942fcf7bee79/galera/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.825284 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_97b35086-f8bc-4035-9685-212989ef3b77/nova-metadata-metadata/0.log" Dec 12 09:41:02 crc kubenswrapper[4867]: I1212 09:41:02.852875 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0583cfff-0ec3-4fad-84de-3c9f6f3dd49c/mysql-bootstrap/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.026391 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0583cfff-0ec3-4fad-84de-3c9f6f3dd49c/mysql-bootstrap/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.038041 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_07f2128b-4e78-430a-9af7-45cf3e402573/openstackclient/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.070159 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0583cfff-0ec3-4fad-84de-3c9f6f3dd49c/galera/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.218383 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c21b182d-1e8c-41aa-8fd3-51bc302238de/openstack-network-exporter/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.244818 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c21b182d-1e8c-41aa-8fd3-51bc302238de/ovn-northd/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.451020 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-rtsj8_f4807fa4-5151-46f0-88ea-d9878b926bae/ovn-openstack-openstack-cell1/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.490984 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f852bd2b-64b9-432a-a921-1f16c6f67cea/ovsdbserver-nb/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.745860 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f852bd2b-64b9-432a-a921-1f16c6f67cea/openstack-network-exporter/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.920116 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_709a4369-8271-49da-b10e-810db17bb0db/openstack-network-exporter/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.969981 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_709a4369-8271-49da-b10e-810db17bb0db/ovsdbserver-nb/0.log" Dec 12 09:41:03 crc kubenswrapper[4867]: I1212 09:41:03.990703 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_dd670fea-d4c3-4a38-91c5-918f6db4fb5b/openstack-network-exporter/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.103287 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_dd670fea-d4c3-4a38-91c5-918f6db4fb5b/ovsdbserver-nb/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.153159 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36c35414-d05a-45e6-92c0-f28e80b01c00/openstack-network-exporter/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.262665 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36c35414-d05a-45e6-92c0-f28e80b01c00/ovsdbserver-sb/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.529331 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_1ddac016-beca-4499-9ac8-2b846fb51a03/openstack-network-exporter/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.529370 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_f98a91b8-7475-452a-a7fe-eafbd824c99f/openstack-network-exporter/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.535096 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_f98a91b8-7475-452a-a7fe-eafbd824c99f/ovsdbserver-sb/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.535407 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_1ddac016-beca-4499-9ac8-2b846fb51a03/ovsdbserver-sb/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.737024 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c7rswk_821d9b97-2dd4-456d-8dbc-5bb0da1685f7/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.790007 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68c67cf9f8-5kb5p_f0286224-fef5-43fc-895c-15806d60a789/placement-api/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.901667 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68c67cf9f8-5kb5p_f0286224-fef5-43fc-895c-15806d60a789/placement-log/0.log" Dec 12 09:41:04 crc kubenswrapper[4867]: I1212 09:41:04.997099 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6d045ee8-6266-4f6f-be38-0a8d39d1fb7e/init-config-reloader/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.199831 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6d045ee8-6266-4f6f-be38-0a8d39d1fb7e/init-config-reloader/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.200481 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6d045ee8-6266-4f6f-be38-0a8d39d1fb7e/config-reloader/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.246349 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6d045ee8-6266-4f6f-be38-0a8d39d1fb7e/thanos-sidecar/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.248859 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_6d045ee8-6266-4f6f-be38-0a8d39d1fb7e/prometheus/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.387483 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ae1a3a41-6283-4343-9d69-3fce5acbf570/setup-container/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.549200 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ae1a3a41-6283-4343-9d69-3fce5acbf570/setup-container/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.568031 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eab4604e-3919-4f11-89ea-057524b8bfb9/setup-container/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.579633 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ae1a3a41-6283-4343-9d69-3fce5acbf570/rabbitmq/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.792171 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eab4604e-3919-4f11-89ea-057524b8bfb9/setup-container/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.845048 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-4dhfn_27e07254-5f63-4c24-996b-178147917e35/reboot-os-openstack-openstack-cell1/0.log" Dec 12 09:41:05 crc kubenswrapper[4867]: I1212 09:41:05.849072 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eab4604e-3919-4f11-89ea-057524b8bfb9/rabbitmq/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.020989 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-76bjl_a6760547-a35d-4676-9ef9-ca5b94a76fb2/run-os-openstack-openstack-cell1/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.078292 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-p46gb_f78f4a6a-78a2-4f00-9835-768392104c41/ssh-known-hosts-openstack/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.274624 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6db4df84b6-rncss_3066d921-d258-4f35-b6fe-2dc9203299e5/proxy-server/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.351344 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6db4df84b6-rncss_3066d921-d258-4f35-b6fe-2dc9203299e5/proxy-httpd/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.367503 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-k2p9r_b54aaf28-10a9-4b58-b752-278929bf0bb8/swift-ring-rebalance/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.499484 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-nnz9m_9cb08d58-7abd-4896-975d-aa575c651851/telemetry-openstack-openstack-cell1/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.663999 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ea049736-19d9-474a-8c1f-39193e6f35df/tempest-tests-tempest-tests-runner/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.697640 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_767fc201-7186-425f-95a3-fef4e63172ed/test-operator-logs-container/0.log" Dec 12 09:41:06 crc kubenswrapper[4867]: I1212 09:41:06.850914 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-bwhwn_55c56e88-ce42-4abb-bdfb-e68fd154fea6/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 12 09:41:07 crc kubenswrapper[4867]: I1212 09:41:07.121271 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-w886f_bd4446dc-714b-4d9f-81f3-c1dcdeb7b8bc/validate-network-openstack-openstack-cell1/0.log" Dec 12 09:41:11 crc kubenswrapper[4867]: I1212 09:41:11.838882 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:41:11 crc kubenswrapper[4867]: E1212 09:41:11.839806 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:41:24 crc kubenswrapper[4867]: I1212 09:41:24.838398 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:41:24 crc kubenswrapper[4867]: E1212 09:41:24.839294 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.441395 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/util/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.595599 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/pull/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.650964 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/util/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.658569 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/pull/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.807755 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/util/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.815718 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/extract/0.log" Dec 12 09:41:32 crc kubenswrapper[4867]: I1212 09:41:32.816571 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_65ebaa037d815792c1ca03416167368b18d08b26771bfdbda948c8e9d1526lw_a40a7208-8353-437c-900b-6341f482a91a/pull/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.001467 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-9rhfd_d774c064-6d06-4d04-8c11-7008260c0044/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.089498 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-5zfsk_8731ddb0-38b0-4de8-895a-c880c8899d26/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.142422 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-9rhfd_d774c064-6d06-4d04-8c11-7008260c0044/manager/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.271244 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-5zfsk_8731ddb0-38b0-4de8-895a-c880c8899d26/manager/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.306448 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-kf2v2_2962c7e5-a1d8-451b-9d65-f4e6e8295e2f/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.382963 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-kf2v2_2962c7e5-a1d8-451b-9d65-f4e6e8295e2f/manager/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.489778 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-5mwnj_528acf70-c2ad-4968-88ad-24531e8f5d12/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.658104 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-5mwnj_528acf70-c2ad-4968-88ad-24531e8f5d12/manager/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.728845 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-qwkr9_09531d04-5bd7-4457-af43-296c2e76cdfb/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.815152 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-qwkr9_09531d04-5bd7-4457-af43-296c2e76cdfb/manager/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.895563 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-z9jcv_93fd7786-fb3e-45aa-92d2-6b6fa38daf56/kube-rbac-proxy/0.log" Dec 12 09:41:33 crc kubenswrapper[4867]: I1212 09:41:33.949984 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-z9jcv_93fd7786-fb3e-45aa-92d2-6b6fa38daf56/manager/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.091032 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-jqlwv_6c74d9ef-6d91-4c62-bfe4-e4811b78adf6/kube-rbac-proxy/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.228176 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-rfkwx_eacc6270-8929-44a1-a971-f2a6b7582103/kube-rbac-proxy/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.330372 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-rfkwx_eacc6270-8929-44a1-a971-f2a6b7582103/manager/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.444863 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-62qkh_9c8a7928-c0c9-4516-8e34-273afc40ccae/kube-rbac-proxy/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.665731 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-2hqbr_eab0a735-13ed-4d91-a4c5-d3c09d99139d/kube-rbac-proxy/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.677168 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-jqlwv_6c74d9ef-6d91-4c62-bfe4-e4811b78adf6/manager/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.699127 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-62qkh_9c8a7928-c0c9-4516-8e34-273afc40ccae/manager/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.822694 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-2hqbr_eab0a735-13ed-4d91-a4c5-d3c09d99139d/manager/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.898031 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-frb6j_f70067b7-bc5e-4f40-9b90-7601ac37aa89/kube-rbac-proxy/0.log" Dec 12 09:41:34 crc kubenswrapper[4867]: I1212 09:41:34.926354 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-frb6j_f70067b7-bc5e-4f40-9b90-7601ac37aa89/manager/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.098509 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-fhp5f_5a7bb06a-0518-4384-b694-b73194df4a05/kube-rbac-proxy/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.173114 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-fhp5f_5a7bb06a-0518-4384-b694-b73194df4a05/manager/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.262765 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-6drsk_09951d5e-0bdd-4a76-b433-880fa1bb7abd/kube-rbac-proxy/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.392377 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c8dlk_b29bfc3e-61ea-443a-af46-13583b7d619f/kube-rbac-proxy/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.433139 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-6drsk_09951d5e-0bdd-4a76-b433-880fa1bb7abd/manager/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.572487 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-c8dlk_b29bfc3e-61ea-443a-af46-13583b7d619f/manager/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.714276 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5c747c4-kzx62_eb70d75b-9337-43a0-960b-ec19d05690bc/kube-rbac-proxy/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.716761 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5c747c4-kzx62_eb70d75b-9337-43a0-960b-ec19d05690bc/manager/0.log" Dec 12 09:41:35 crc kubenswrapper[4867]: I1212 09:41:35.837870 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:41:35 crc kubenswrapper[4867]: E1212 09:41:35.838301 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.090435 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-799b66f579-q8vtt_98b6024c-e03c-45a1-856b-e3f4af5483cd/operator/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.364909 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5zjf8_42800642-ed16-4b36-8587-8d136d065283/kube-rbac-proxy/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.376178 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-qz9fj_c06744d8-3f13-4940-9878-f2bf5a1eb5ee/registry-server/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.517006 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5zjf8_42800642-ed16-4b36-8587-8d136d065283/manager/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.609701 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vhfhv_86d48a0c-c93a-4090-8f13-f9d6ec122a68/kube-rbac-proxy/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.672137 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vhfhv_86d48a0c-c93a-4090-8f13-f9d6ec122a68/manager/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.844278 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-7rkgr_08998d49-f3b6-4849-8182-8199f1f5c4c6/operator/0.log" Dec 12 09:41:36 crc kubenswrapper[4867]: I1212 09:41:36.905025 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-r24lf_b63f08d0-fff4-4852-9cd2-6467bf523366/kube-rbac-proxy/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.053801 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-r24lf_b63f08d0-fff4-4852-9cd2-6467bf523366/manager/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.157695 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-q84bc_e17f7472-9f06-46f7-a911-96a88b93d30e/kube-rbac-proxy/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.286351 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4cfxw_11ecafd2-343b-4ee3-9ffa-c6987230ccc3/kube-rbac-proxy/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.417087 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-4cfxw_11ecafd2-343b-4ee3-9ffa-c6987230ccc3/manager/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.548438 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-q84bc_e17f7472-9f06-46f7-a911-96a88b93d30e/manager/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.816167 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-vnbsx_a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04/kube-rbac-proxy/0.log" Dec 12 09:41:37 crc kubenswrapper[4867]: I1212 09:41:37.879389 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-vnbsx_a01f2b25-dab2-4d4d-8d6a-ba3249ec2f04/manager/0.log" Dec 12 09:41:38 crc kubenswrapper[4867]: I1212 09:41:38.528926 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5b74fbd87-zjj58_9a77c831-df7a-4167-a8d4-5ca630bbefb0/manager/0.log" Dec 12 09:41:49 crc kubenswrapper[4867]: I1212 09:41:49.838386 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:41:49 crc kubenswrapper[4867]: E1212 09:41:49.839147 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:41:55 crc kubenswrapper[4867]: I1212 09:41:55.520784 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-n64dp_3d58b26a-54e5-4699-9a2c-e4af556a4405/control-plane-machine-set-operator/0.log" Dec 12 09:41:55 crc kubenswrapper[4867]: I1212 09:41:55.709679 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7zlz5_b9ce7d0e-4827-492c-a71e-53299557a2b5/machine-api-operator/0.log" Dec 12 09:41:55 crc kubenswrapper[4867]: I1212 09:41:55.712038 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7zlz5_b9ce7d0e-4827-492c-a71e-53299557a2b5/kube-rbac-proxy/0.log" Dec 12 09:42:02 crc kubenswrapper[4867]: I1212 09:42:02.844935 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:42:02 crc kubenswrapper[4867]: E1212 09:42:02.846899 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:42:07 crc kubenswrapper[4867]: I1212 09:42:07.103245 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-jm9f4_3c8d2ab5-ca42-4bad-96fd-efa1b292cf70/cert-manager-controller/0.log" Dec 12 09:42:07 crc kubenswrapper[4867]: I1212 09:42:07.228440 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-7ll2f_492804a0-56c4-44eb-b9a6-1328cd8b7e84/cert-manager-cainjector/0.log" Dec 12 09:42:07 crc kubenswrapper[4867]: I1212 09:42:07.233297 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-l4pl6_11283469-26a6-4a73-a330-e3427e0b22b1/cert-manager-webhook/0.log" Dec 12 09:42:17 crc kubenswrapper[4867]: I1212 09:42:17.837773 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:42:17 crc kubenswrapper[4867]: E1212 09:42:17.838478 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.456393 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-22djm_1ce76a66-a552-43f5-b37e-d62306ce2422/nmstate-console-plugin/0.log" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.610846 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-wzr2t_ed16d134-4ec5-44bc-bb0b-5281bea6dad3/nmstate-handler/0.log" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.645478 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-2jnfh_f4cebb09-194f-4459-b31e-43fdbd632cc1/kube-rbac-proxy/0.log" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.647128 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-2jnfh_f4cebb09-194f-4459-b31e-43fdbd632cc1/nmstate-metrics/0.log" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.830369 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-ts9zq_01d83fb8-4fb3-43f4-a597-daed3e63f0bb/nmstate-operator/0.log" Dec 12 09:42:18 crc kubenswrapper[4867]: I1212 09:42:18.882967 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-c7b78_444fb8e2-d658-4b0f-b572-d6acbfb93fcd/nmstate-webhook/0.log" Dec 12 09:42:30 crc kubenswrapper[4867]: I1212 09:42:30.847966 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:42:30 crc kubenswrapper[4867]: E1212 09:42:30.849066 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.156057 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-9wsvj_d21a5490-c13d-4182-a608-81bfa11f5327/kube-rbac-proxy/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.400120 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-frr-files/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.619605 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-9wsvj_d21a5490-c13d-4182-a608-81bfa11f5327/controller/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.677412 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-metrics/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.677896 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-frr-files/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.711052 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-reloader/0.log" Dec 12 09:42:33 crc kubenswrapper[4867]: I1212 09:42:33.832589 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-reloader/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.011569 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-frr-files/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.049488 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-metrics/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.060895 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-reloader/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.078349 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-metrics/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.217900 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-reloader/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.247541 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-frr-files/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.263507 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/controller/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.292691 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/cp-metrics/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.503148 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/kube-rbac-proxy-frr/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.515635 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/frr-metrics/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.546621 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/kube-rbac-proxy/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.709356 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/reloader/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.779419 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-fbwh6_dbbdf628-904d-4b3a-b07f-40f46b5a57b2/frr-k8s-webhook-server/0.log" Dec 12 09:42:34 crc kubenswrapper[4867]: I1212 09:42:34.993449 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-85cd4f79db-fjxlm_2bdc1624-a89c-4703-9d6b-12dd126ca87f/manager/0.log" Dec 12 09:42:35 crc kubenswrapper[4867]: I1212 09:42:35.164415 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-97cbb5d69-6gxhd_2736a731-8ce4-4551-9fec-190c75ef0f04/webhook-server/0.log" Dec 12 09:42:35 crc kubenswrapper[4867]: I1212 09:42:35.348756 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cwrkp_68b218be-4b27-4caa-8f64-af253c7a2dff/kube-rbac-proxy/0.log" Dec 12 09:42:36 crc kubenswrapper[4867]: I1212 09:42:36.239030 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-cwrkp_68b218be-4b27-4caa-8f64-af253c7a2dff/speaker/0.log" Dec 12 09:42:37 crc kubenswrapper[4867]: I1212 09:42:37.755320 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-4rx5r_45a86044-479c-496a-bce8-8a8b61199127/frr/0.log" Dec 12 09:42:44 crc kubenswrapper[4867]: I1212 09:42:44.838951 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:42:44 crc kubenswrapper[4867]: E1212 09:42:44.839745 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.493787 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/util/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.675869 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/util/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.705002 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/pull/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.707372 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/pull/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.923211 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/extract/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.930384 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/util/0.log" Dec 12 09:42:49 crc kubenswrapper[4867]: I1212 09:42:49.953667 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931at7f55_b966f81e-701c-49f5-bedf-732815bc91bb/pull/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.094061 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/util/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.341255 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/util/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.349445 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/pull/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.374257 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/pull/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.602676 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/util/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.626805 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/extract/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.659932 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d47v5q6_ce7963a0-6daf-4995-b782-f6d03ee9f1e7/pull/0.log" Dec 12 09:42:50 crc kubenswrapper[4867]: I1212 09:42:50.824163 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/util/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.060061 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/pull/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.071122 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/util/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.077718 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/pull/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.244172 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/pull/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.307868 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/extract/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.342650 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210fp5th_a27475d4-d4ab-4f25-b83c-d107efa67cd7/util/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.667612 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/util/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.914842 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/pull/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.924854 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/util/0.log" Dec 12 09:42:51 crc kubenswrapper[4867]: I1212 09:42:51.962312 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/pull/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.149531 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/extract/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.279176 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/pull/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.285580 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8gnqmb_3951e0b1-ac70-4fe4-b9ec-2c33ce2bba08/util/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.368613 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-utilities/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.549717 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-utilities/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.578561 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-content/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.585761 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-content/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.802834 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-utilities/0.log" Dec 12 09:42:52 crc kubenswrapper[4867]: I1212 09:42:52.835098 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/extract-content/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.030334 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-utilities/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.355484 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-content/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.364121 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-utilities/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.406195 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-content/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.613397 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-content/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.675555 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/extract-utilities/0.log" Dec 12 09:42:53 crc kubenswrapper[4867]: I1212 09:42:53.894847 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bb7j7_d6e1399d-7725-4ea9-805d-1a363cc0bd09/marketplace-operator/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.112425 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-utilities/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.373897 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-content/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.417469 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-content/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.469585 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-utilities/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.609458 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-p9dng_097fc828-65da-4d60-9aac-3ff18d583b12/registry-server/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.675371 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-utilities/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.737200 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/extract-content/0.log" Dec 12 09:42:54 crc kubenswrapper[4867]: I1212 09:42:54.923086 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-utilities/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.091676 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2rmpq_715ca56a-e74a-4ca4-a30c-7779d73b4fb4/registry-server/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.122251 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-utilities/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.122436 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-content/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.178788 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-content/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.285133 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqqrr_b08a5662-0455-4b9e-84db-f471b8a74d33/registry-server/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.390848 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-utilities/0.log" Dec 12 09:42:55 crc kubenswrapper[4867]: I1212 09:42:55.495198 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/extract-content/0.log" Dec 12 09:42:56 crc kubenswrapper[4867]: I1212 09:42:56.643370 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2hql5_196dd20e-43b3-4f25-a61c-a94d958bb555/registry-server/0.log" Dec 12 09:42:57 crc kubenswrapper[4867]: I1212 09:42:57.867324 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:42:57 crc kubenswrapper[4867]: E1212 09:42:57.867746 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.699874 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:08 crc kubenswrapper[4867]: E1212 09:43:08.700871 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="extract-content" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.700885 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="extract-content" Dec 12 09:43:08 crc kubenswrapper[4867]: E1212 09:43:08.700894 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="registry-server" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.700900 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="registry-server" Dec 12 09:43:08 crc kubenswrapper[4867]: E1212 09:43:08.700909 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="extract-utilities" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.700915 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="extract-utilities" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.701153 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="e931a3c8-10b2-4cb9-8145-e9111a6da0a6" containerName="registry-server" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.703417 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.712893 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.787453 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.787849 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.787894 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92svv\" (UniqueName: \"kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.889869 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.889946 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.890008 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92svv\" (UniqueName: \"kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.890595 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.890642 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:08 crc kubenswrapper[4867]: I1212 09:43:08.917705 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92svv\" (UniqueName: \"kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv\") pod \"redhat-operators-ggxkr\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.025526 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.247412 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-df5xh_751ded8c-1d36-4275-a0b5-d6444285b50b/prometheus-operator/0.log" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.494436 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-98568f549-k4lz7_b9258b12-9a2e-4219-aaa4-1ec3717692b2/prometheus-operator-admission-webhook/0.log" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.619157 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-98568f549-rfr5n_3bc2cd37-97e9-43b8-a499-0b706dcc7fd6/prometheus-operator-admission-webhook/0.log" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.653758 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.801804 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-56mq5_035114f3-82c3-4566-83ab-7b4151b7232d/operator/0.log" Dec 12 09:43:09 crc kubenswrapper[4867]: I1212 09:43:09.921962 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-kgxfk_545d1312-b5d7-40fa-85c4-d54c8e09c7a2/perses-operator/0.log" Dec 12 09:43:10 crc kubenswrapper[4867]: I1212 09:43:10.133947 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerStarted","Data":"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af"} Dec 12 09:43:10 crc kubenswrapper[4867]: I1212 09:43:10.134000 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerStarted","Data":"e106dab50ddd535898a14e5b406ee99f2ddfb872f23e2a540ee477005107c06d"} Dec 12 09:43:11 crc kubenswrapper[4867]: I1212 09:43:11.150866 4867 generic.go:334] "Generic (PLEG): container finished" podID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerID="be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af" exitCode=0 Dec 12 09:43:11 crc kubenswrapper[4867]: I1212 09:43:11.151310 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerDied","Data":"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af"} Dec 12 09:43:11 crc kubenswrapper[4867]: I1212 09:43:11.162785 4867 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 12 09:43:12 crc kubenswrapper[4867]: I1212 09:43:12.162597 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerStarted","Data":"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f"} Dec 12 09:43:12 crc kubenswrapper[4867]: I1212 09:43:12.846890 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:43:12 crc kubenswrapper[4867]: E1212 09:43:12.847213 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:43:15 crc kubenswrapper[4867]: I1212 09:43:15.197318 4867 generic.go:334] "Generic (PLEG): container finished" podID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerID="e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f" exitCode=0 Dec 12 09:43:15 crc kubenswrapper[4867]: I1212 09:43:15.197391 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerDied","Data":"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f"} Dec 12 09:43:18 crc kubenswrapper[4867]: I1212 09:43:18.230222 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerStarted","Data":"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59"} Dec 12 09:43:18 crc kubenswrapper[4867]: I1212 09:43:18.252653 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ggxkr" podStartSLOduration=4.197683287 podStartE2EDuration="10.252634284s" podCreationTimestamp="2025-12-12 09:43:08 +0000 UTC" firstStartedPulling="2025-12-12 09:43:11.162467132 +0000 UTC m=+10478.733848411" lastFinishedPulling="2025-12-12 09:43:17.217418139 +0000 UTC m=+10484.788799408" observedRunningTime="2025-12-12 09:43:18.248064071 +0000 UTC m=+10485.819445340" watchObservedRunningTime="2025-12-12 09:43:18.252634284 +0000 UTC m=+10485.824015553" Dec 12 09:43:19 crc kubenswrapper[4867]: I1212 09:43:19.026633 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:19 crc kubenswrapper[4867]: I1212 09:43:19.026689 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:20 crc kubenswrapper[4867]: I1212 09:43:20.081922 4867 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ggxkr" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="registry-server" probeResult="failure" output=< Dec 12 09:43:20 crc kubenswrapper[4867]: timeout: failed to connect service ":50051" within 1s Dec 12 09:43:20 crc kubenswrapper[4867]: > Dec 12 09:43:26 crc kubenswrapper[4867]: I1212 09:43:26.838811 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:43:26 crc kubenswrapper[4867]: E1212 09:43:26.839561 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:43:29 crc kubenswrapper[4867]: I1212 09:43:29.088835 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:29 crc kubenswrapper[4867]: I1212 09:43:29.183520 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:29 crc kubenswrapper[4867]: I1212 09:43:29.331734 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:30 crc kubenswrapper[4867]: I1212 09:43:30.363375 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ggxkr" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="registry-server" containerID="cri-o://a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59" gracePeriod=2 Dec 12 09:43:30 crc kubenswrapper[4867]: I1212 09:43:30.981025 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.120257 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content\") pod \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.120337 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92svv\" (UniqueName: \"kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv\") pod \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.120377 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities\") pod \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\" (UID: \"a23ce8cf-91f9-443d-a17d-bad59f4f6c11\") " Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.121104 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities" (OuterVolumeSpecName: "utilities") pod "a23ce8cf-91f9-443d-a17d-bad59f4f6c11" (UID: "a23ce8cf-91f9-443d-a17d-bad59f4f6c11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.127442 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv" (OuterVolumeSpecName: "kube-api-access-92svv") pod "a23ce8cf-91f9-443d-a17d-bad59f4f6c11" (UID: "a23ce8cf-91f9-443d-a17d-bad59f4f6c11"). InnerVolumeSpecName "kube-api-access-92svv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.222888 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92svv\" (UniqueName: \"kubernetes.io/projected/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-kube-api-access-92svv\") on node \"crc\" DevicePath \"\"" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.222927 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.247339 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a23ce8cf-91f9-443d-a17d-bad59f4f6c11" (UID: "a23ce8cf-91f9-443d-a17d-bad59f4f6c11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.325090 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a23ce8cf-91f9-443d-a17d-bad59f4f6c11-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.375741 4867 generic.go:334] "Generic (PLEG): container finished" podID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerID="a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59" exitCode=0 Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.375783 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerDied","Data":"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59"} Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.375806 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggxkr" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.375825 4867 scope.go:117] "RemoveContainer" containerID="a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.375811 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggxkr" event={"ID":"a23ce8cf-91f9-443d-a17d-bad59f4f6c11","Type":"ContainerDied","Data":"e106dab50ddd535898a14e5b406ee99f2ddfb872f23e2a540ee477005107c06d"} Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.395904 4867 scope.go:117] "RemoveContainer" containerID="e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.418640 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.430644 4867 scope.go:117] "RemoveContainer" containerID="be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.435357 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ggxkr"] Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.485946 4867 scope.go:117] "RemoveContainer" containerID="a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59" Dec 12 09:43:31 crc kubenswrapper[4867]: E1212 09:43:31.486892 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59\": container with ID starting with a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59 not found: ID does not exist" containerID="a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.487030 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59"} err="failed to get container status \"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59\": rpc error: code = NotFound desc = could not find container \"a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59\": container with ID starting with a82d214dd10f99bbe6b38aafc7d91d57183233084d751a7a5f6740a1be5efb59 not found: ID does not exist" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.487141 4867 scope.go:117] "RemoveContainer" containerID="e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f" Dec 12 09:43:31 crc kubenswrapper[4867]: E1212 09:43:31.487880 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f\": container with ID starting with e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f not found: ID does not exist" containerID="e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.488005 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f"} err="failed to get container status \"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f\": rpc error: code = NotFound desc = could not find container \"e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f\": container with ID starting with e9b07c8a2cd4b2b33cac8c5f7727196a29802754e187e5c649e2d32b90aaed8f not found: ID does not exist" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.488115 4867 scope.go:117] "RemoveContainer" containerID="be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af" Dec 12 09:43:31 crc kubenswrapper[4867]: E1212 09:43:31.488685 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af\": container with ID starting with be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af not found: ID does not exist" containerID="be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af" Dec 12 09:43:31 crc kubenswrapper[4867]: I1212 09:43:31.488829 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af"} err="failed to get container status \"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af\": rpc error: code = NotFound desc = could not find container \"be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af\": container with ID starting with be9352827483d5638d78acfb6671fb7fedb63a431e92394fac1030c2f41640af not found: ID does not exist" Dec 12 09:43:32 crc kubenswrapper[4867]: I1212 09:43:32.964162 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" path="/var/lib/kubelet/pods/a23ce8cf-91f9-443d-a17d-bad59f4f6c11/volumes" Dec 12 09:43:37 crc kubenswrapper[4867]: I1212 09:43:37.841696 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:43:37 crc kubenswrapper[4867]: E1212 09:43:37.843453 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:43:48 crc kubenswrapper[4867]: I1212 09:43:48.838775 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:43:48 crc kubenswrapper[4867]: E1212 09:43:48.839928 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:43:59 crc kubenswrapper[4867]: I1212 09:43:59.838183 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:43:59 crc kubenswrapper[4867]: E1212 09:43:59.839123 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:44:07 crc kubenswrapper[4867]: I1212 09:44:07.649313 4867 scope.go:117] "RemoveContainer" containerID="43ef36457cde94731bebf7f8d819341eccbb31dc0b062eb533ebb93129ebafa6" Dec 12 09:44:12 crc kubenswrapper[4867]: I1212 09:44:12.844406 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:44:12 crc kubenswrapper[4867]: E1212 09:44:12.845128 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:44:24 crc kubenswrapper[4867]: I1212 09:44:24.842138 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:44:24 crc kubenswrapper[4867]: E1212 09:44:24.842992 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:44:36 crc kubenswrapper[4867]: I1212 09:44:36.838959 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:44:36 crc kubenswrapper[4867]: E1212 09:44:36.839766 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:44:52 crc kubenswrapper[4867]: I1212 09:44:52.851960 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:44:52 crc kubenswrapper[4867]: E1212 09:44:52.852762 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.172817 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8"] Dec 12 09:45:00 crc kubenswrapper[4867]: E1212 09:45:00.173892 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="extract-utilities" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.173911 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="extract-utilities" Dec 12 09:45:00 crc kubenswrapper[4867]: E1212 09:45:00.173932 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="extract-content" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.173938 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="extract-content" Dec 12 09:45:00 crc kubenswrapper[4867]: E1212 09:45:00.173965 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="registry-server" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.173971 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="registry-server" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.174190 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="a23ce8cf-91f9-443d-a17d-bad59f4f6c11" containerName="registry-server" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.175824 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.177707 4867 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.178041 4867 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.190683 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srv2z\" (UniqueName: \"kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.190768 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.191036 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.207329 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8"] Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.292588 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srv2z\" (UniqueName: \"kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.292667 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.292812 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.293832 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.301065 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.312252 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srv2z\" (UniqueName: \"kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z\") pod \"collect-profiles-29425545-62ss8\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.505222 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:00 crc kubenswrapper[4867]: I1212 09:45:00.988455 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8"] Dec 12 09:45:01 crc kubenswrapper[4867]: I1212 09:45:01.217893 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" event={"ID":"f54356a5-cdc9-4732-8ae4-773aa3d32a30","Type":"ContainerStarted","Data":"5d922a6e2ddfd71e0dcdd2449b883cf23570c92fd2dafa77ef3791258525db4f"} Dec 12 09:45:01 crc kubenswrapper[4867]: I1212 09:45:01.218311 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" event={"ID":"f54356a5-cdc9-4732-8ae4-773aa3d32a30","Type":"ContainerStarted","Data":"4aaaca3bf9c263e14212ffdb33140d636146fe81df3f02368066d03a0782e93b"} Dec 12 09:45:01 crc kubenswrapper[4867]: I1212 09:45:01.237010 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" podStartSLOduration=1.236968515 podStartE2EDuration="1.236968515s" podCreationTimestamp="2025-12-12 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-12 09:45:01.232132664 +0000 UTC m=+10588.803513933" watchObservedRunningTime="2025-12-12 09:45:01.236968515 +0000 UTC m=+10588.808349784" Dec 12 09:45:02 crc kubenswrapper[4867]: I1212 09:45:02.229921 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" event={"ID":"f54356a5-cdc9-4732-8ae4-773aa3d32a30","Type":"ContainerDied","Data":"5d922a6e2ddfd71e0dcdd2449b883cf23570c92fd2dafa77ef3791258525db4f"} Dec 12 09:45:02 crc kubenswrapper[4867]: I1212 09:45:02.229755 4867 generic.go:334] "Generic (PLEG): container finished" podID="f54356a5-cdc9-4732-8ae4-773aa3d32a30" containerID="5d922a6e2ddfd71e0dcdd2449b883cf23570c92fd2dafa77ef3791258525db4f" exitCode=0 Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.595265 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.760570 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume\") pod \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.760635 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume\") pod \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.760698 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srv2z\" (UniqueName: \"kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z\") pod \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\" (UID: \"f54356a5-cdc9-4732-8ae4-773aa3d32a30\") " Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.762185 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume" (OuterVolumeSpecName: "config-volume") pod "f54356a5-cdc9-4732-8ae4-773aa3d32a30" (UID: "f54356a5-cdc9-4732-8ae4-773aa3d32a30"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.767191 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z" (OuterVolumeSpecName: "kube-api-access-srv2z") pod "f54356a5-cdc9-4732-8ae4-773aa3d32a30" (UID: "f54356a5-cdc9-4732-8ae4-773aa3d32a30"). InnerVolumeSpecName "kube-api-access-srv2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.783404 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f54356a5-cdc9-4732-8ae4-773aa3d32a30" (UID: "f54356a5-cdc9-4732-8ae4-773aa3d32a30"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.864234 4867 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f54356a5-cdc9-4732-8ae4-773aa3d32a30-config-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.864299 4867 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f54356a5-cdc9-4732-8ae4-773aa3d32a30-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:03 crc kubenswrapper[4867]: I1212 09:45:03.864311 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srv2z\" (UniqueName: \"kubernetes.io/projected/f54356a5-cdc9-4732-8ae4-773aa3d32a30-kube-api-access-srv2z\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.255776 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" event={"ID":"f54356a5-cdc9-4732-8ae4-773aa3d32a30","Type":"ContainerDied","Data":"4aaaca3bf9c263e14212ffdb33140d636146fe81df3f02368066d03a0782e93b"} Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.256236 4867 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aaaca3bf9c263e14212ffdb33140d636146fe81df3f02368066d03a0782e93b" Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.255828 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29425545-62ss8" Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.321532 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl"] Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.332591 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29425500-552gl"] Dec 12 09:45:04 crc kubenswrapper[4867]: I1212 09:45:04.852329 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dddadd53-c9e0-4524-a315-72157150ca2c" path="/var/lib/kubelet/pods/dddadd53-c9e0-4524-a315-72157150ca2c/volumes" Dec 12 09:45:05 crc kubenswrapper[4867]: I1212 09:45:05.837968 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:45:05 crc kubenswrapper[4867]: E1212 09:45:05.838334 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:45:07 crc kubenswrapper[4867]: I1212 09:45:07.736062 4867 scope.go:117] "RemoveContainer" containerID="ac603494517619aefe6e568808eb21939b2e39e233087b6bf515146685bf60f5" Dec 12 09:45:07 crc kubenswrapper[4867]: I1212 09:45:07.757542 4867 scope.go:117] "RemoveContainer" containerID="3225e49e590bccf5d41b9b0cef876c171c4a3e01ce802a7ac3feb3ad903a7a47" Dec 12 09:45:18 crc kubenswrapper[4867]: I1212 09:45:18.838781 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:45:18 crc kubenswrapper[4867]: E1212 09:45:18.839688 4867 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-q8lbh_openshift-machine-config-operator(f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae)\"" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" Dec 12 09:45:25 crc kubenswrapper[4867]: I1212 09:45:25.466666 4867 generic.go:334] "Generic (PLEG): container finished" podID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerID="9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2" exitCode=0 Dec 12 09:45:25 crc kubenswrapper[4867]: I1212 09:45:25.466774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ksppc/must-gather-r6cx8" event={"ID":"1b51416a-1117-4f1c-8d12-b9fdc1c1307a","Type":"ContainerDied","Data":"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2"} Dec 12 09:45:25 crc kubenswrapper[4867]: I1212 09:45:25.467725 4867 scope.go:117] "RemoveContainer" containerID="9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2" Dec 12 09:45:25 crc kubenswrapper[4867]: I1212 09:45:25.863531 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksppc_must-gather-r6cx8_1b51416a-1117-4f1c-8d12-b9fdc1c1307a/gather/0.log" Dec 12 09:45:30 crc kubenswrapper[4867]: I1212 09:45:30.838423 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" Dec 12 09:45:31 crc kubenswrapper[4867]: I1212 09:45:31.524439 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"f9486a6274c2ad06071243c6ef7c9e72e8c56f8224e3124bdf4595856442807f"} Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.000062 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ksppc/must-gather-r6cx8"] Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.000920 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ksppc/must-gather-r6cx8" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="copy" containerID="cri-o://a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da" gracePeriod=2 Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.012981 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ksppc/must-gather-r6cx8"] Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.423338 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksppc_must-gather-r6cx8_1b51416a-1117-4f1c-8d12-b9fdc1c1307a/copy/0.log" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.424802 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.547139 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output\") pod \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.547333 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxd7s\" (UniqueName: \"kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s\") pod \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\" (UID: \"1b51416a-1117-4f1c-8d12-b9fdc1c1307a\") " Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.552782 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s" (OuterVolumeSpecName: "kube-api-access-bxd7s") pod "1b51416a-1117-4f1c-8d12-b9fdc1c1307a" (UID: "1b51416a-1117-4f1c-8d12-b9fdc1c1307a"). InnerVolumeSpecName "kube-api-access-bxd7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.562387 4867 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ksppc_must-gather-r6cx8_1b51416a-1117-4f1c-8d12-b9fdc1c1307a/copy/0.log" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.562923 4867 generic.go:334] "Generic (PLEG): container finished" podID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerID="a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da" exitCode=143 Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.562969 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ksppc/must-gather-r6cx8" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.562991 4867 scope.go:117] "RemoveContainer" containerID="a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.649315 4867 scope.go:117] "RemoveContainer" containerID="9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.649648 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxd7s\" (UniqueName: \"kubernetes.io/projected/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-kube-api-access-bxd7s\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.744117 4867 scope.go:117] "RemoveContainer" containerID="a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da" Dec 12 09:45:35 crc kubenswrapper[4867]: E1212 09:45:35.745012 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da\": container with ID starting with a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da not found: ID does not exist" containerID="a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.745046 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da"} err="failed to get container status \"a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da\": rpc error: code = NotFound desc = could not find container \"a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da\": container with ID starting with a8b65d0660ec488deaa063804d6c9bcf4d014b8307e4e982589d26a9415568da not found: ID does not exist" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.745072 4867 scope.go:117] "RemoveContainer" containerID="9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2" Dec 12 09:45:35 crc kubenswrapper[4867]: E1212 09:45:35.745351 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2\": container with ID starting with 9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2 not found: ID does not exist" containerID="9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.745373 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2"} err="failed to get container status \"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2\": rpc error: code = NotFound desc = could not find container \"9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2\": container with ID starting with 9d01a59377d9dad8a69026fbbe3883dc145689d1c11d75159fe5f0c0683782b2 not found: ID does not exist" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.783547 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1b51416a-1117-4f1c-8d12-b9fdc1c1307a" (UID: "1b51416a-1117-4f1c-8d12-b9fdc1c1307a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:45:35 crc kubenswrapper[4867]: I1212 09:45:35.852366 4867 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1b51416a-1117-4f1c-8d12-b9fdc1c1307a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:36 crc kubenswrapper[4867]: I1212 09:45:36.852292 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" path="/var/lib/kubelet/pods/1b51416a-1117-4f1c-8d12-b9fdc1c1307a/volumes" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.290436 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:41 crc kubenswrapper[4867]: E1212 09:45:41.293026 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="gather" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293122 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="gather" Dec 12 09:45:41 crc kubenswrapper[4867]: E1212 09:45:41.293226 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f54356a5-cdc9-4732-8ae4-773aa3d32a30" containerName="collect-profiles" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293320 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="f54356a5-cdc9-4732-8ae4-773aa3d32a30" containerName="collect-profiles" Dec 12 09:45:41 crc kubenswrapper[4867]: E1212 09:45:41.293392 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="copy" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293447 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="copy" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293765 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="copy" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293848 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="f54356a5-cdc9-4732-8ae4-773aa3d32a30" containerName="collect-profiles" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.293916 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b51416a-1117-4f1c-8d12-b9fdc1c1307a" containerName="gather" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.295665 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.312124 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.480885 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.481397 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdm5\" (UniqueName: \"kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.481455 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.582926 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdm5\" (UniqueName: \"kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.583026 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.583147 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.583639 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.583664 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.607193 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdm5\" (UniqueName: \"kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5\") pod \"redhat-marketplace-sk4rl\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:41 crc kubenswrapper[4867]: I1212 09:45:41.620653 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:42 crc kubenswrapper[4867]: I1212 09:45:42.119762 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:42 crc kubenswrapper[4867]: I1212 09:45:42.630120 4867 generic.go:334] "Generic (PLEG): container finished" podID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerID="0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560" exitCode=0 Dec 12 09:45:42 crc kubenswrapper[4867]: I1212 09:45:42.630191 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerDied","Data":"0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560"} Dec 12 09:45:42 crc kubenswrapper[4867]: I1212 09:45:42.630421 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerStarted","Data":"76530e68fb4d77f8be76b1fefb853bb79a23a380dd367d30e16d08fbd1fe0db3"} Dec 12 09:45:43 crc kubenswrapper[4867]: I1212 09:45:43.642065 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerStarted","Data":"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d"} Dec 12 09:45:44 crc kubenswrapper[4867]: I1212 09:45:44.652765 4867 generic.go:334] "Generic (PLEG): container finished" podID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerID="cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d" exitCode=0 Dec 12 09:45:44 crc kubenswrapper[4867]: I1212 09:45:44.652814 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerDied","Data":"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d"} Dec 12 09:45:45 crc kubenswrapper[4867]: I1212 09:45:45.666127 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerStarted","Data":"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859"} Dec 12 09:45:45 crc kubenswrapper[4867]: I1212 09:45:45.684122 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sk4rl" podStartSLOduration=2.233420888 podStartE2EDuration="4.684106096s" podCreationTimestamp="2025-12-12 09:45:41 +0000 UTC" firstStartedPulling="2025-12-12 09:45:42.632578491 +0000 UTC m=+10630.203959760" lastFinishedPulling="2025-12-12 09:45:45.083263699 +0000 UTC m=+10632.654644968" observedRunningTime="2025-12-12 09:45:45.683292726 +0000 UTC m=+10633.254674015" watchObservedRunningTime="2025-12-12 09:45:45.684106096 +0000 UTC m=+10633.255487365" Dec 12 09:45:51 crc kubenswrapper[4867]: I1212 09:45:51.620798 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:51 crc kubenswrapper[4867]: I1212 09:45:51.621368 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:51 crc kubenswrapper[4867]: I1212 09:45:51.682493 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:51 crc kubenswrapper[4867]: I1212 09:45:51.854854 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:51 crc kubenswrapper[4867]: I1212 09:45:51.951604 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:53 crc kubenswrapper[4867]: I1212 09:45:53.763285 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sk4rl" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="registry-server" containerID="cri-o://7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859" gracePeriod=2 Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.273182 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.347502 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kdm5\" (UniqueName: \"kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5\") pod \"700228d2-601e-4f5a-be91-d36f5ac39ab5\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.347662 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content\") pod \"700228d2-601e-4f5a-be91-d36f5ac39ab5\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.347780 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities\") pod \"700228d2-601e-4f5a-be91-d36f5ac39ab5\" (UID: \"700228d2-601e-4f5a-be91-d36f5ac39ab5\") " Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.348719 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities" (OuterVolumeSpecName: "utilities") pod "700228d2-601e-4f5a-be91-d36f5ac39ab5" (UID: "700228d2-601e-4f5a-be91-d36f5ac39ab5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.362199 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5" (OuterVolumeSpecName: "kube-api-access-5kdm5") pod "700228d2-601e-4f5a-be91-d36f5ac39ab5" (UID: "700228d2-601e-4f5a-be91-d36f5ac39ab5"). InnerVolumeSpecName "kube-api-access-5kdm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.372445 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "700228d2-601e-4f5a-be91-d36f5ac39ab5" (UID: "700228d2-601e-4f5a-be91-d36f5ac39ab5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.450402 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kdm5\" (UniqueName: \"kubernetes.io/projected/700228d2-601e-4f5a-be91-d36f5ac39ab5-kube-api-access-5kdm5\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.450437 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.450448 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/700228d2-601e-4f5a-be91-d36f5ac39ab5-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.775728 4867 generic.go:334] "Generic (PLEG): container finished" podID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerID="7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859" exitCode=0 Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.775774 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerDied","Data":"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859"} Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.775821 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk4rl" event={"ID":"700228d2-601e-4f5a-be91-d36f5ac39ab5","Type":"ContainerDied","Data":"76530e68fb4d77f8be76b1fefb853bb79a23a380dd367d30e16d08fbd1fe0db3"} Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.775842 4867 scope.go:117] "RemoveContainer" containerID="7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.775843 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk4rl" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.804550 4867 scope.go:117] "RemoveContainer" containerID="cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.831033 4867 scope.go:117] "RemoveContainer" containerID="0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.837138 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.852129 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk4rl"] Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.882761 4867 scope.go:117] "RemoveContainer" containerID="7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859" Dec 12 09:45:54 crc kubenswrapper[4867]: E1212 09:45:54.883191 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859\": container with ID starting with 7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859 not found: ID does not exist" containerID="7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.883256 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859"} err="failed to get container status \"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859\": rpc error: code = NotFound desc = could not find container \"7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859\": container with ID starting with 7aad55e887e46972c8d0f897f266877dfdb65bbd9487f677e8d06abf0a741859 not found: ID does not exist" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.883281 4867 scope.go:117] "RemoveContainer" containerID="cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d" Dec 12 09:45:54 crc kubenswrapper[4867]: E1212 09:45:54.883580 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d\": container with ID starting with cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d not found: ID does not exist" containerID="cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.883638 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d"} err="failed to get container status \"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d\": rpc error: code = NotFound desc = could not find container \"cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d\": container with ID starting with cb57cec25cbd650520296a44d53937d901dbdc85c29d8a75fd3a299d7e480a1d not found: ID does not exist" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.883662 4867 scope.go:117] "RemoveContainer" containerID="0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560" Dec 12 09:45:54 crc kubenswrapper[4867]: E1212 09:45:54.883884 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560\": container with ID starting with 0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560 not found: ID does not exist" containerID="0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560" Dec 12 09:45:54 crc kubenswrapper[4867]: I1212 09:45:54.883912 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560"} err="failed to get container status \"0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560\": rpc error: code = NotFound desc = could not find container \"0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560\": container with ID starting with 0d5d82d27f539ff7541669fa84bfbcee3018ffca6b02a0cc26601b6511419560 not found: ID does not exist" Dec 12 09:45:56 crc kubenswrapper[4867]: I1212 09:45:56.852995 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" path="/var/lib/kubelet/pods/700228d2-601e-4f5a-be91-d36f5ac39ab5/volumes" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.862565 4867 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:26 crc kubenswrapper[4867]: E1212 09:46:26.863283 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="registry-server" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.863294 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="registry-server" Dec 12 09:46:26 crc kubenswrapper[4867]: E1212 09:46:26.863312 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="extract-utilities" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.863318 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="extract-utilities" Dec 12 09:46:26 crc kubenswrapper[4867]: E1212 09:46:26.863338 4867 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="extract-content" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.863343 4867 state_mem.go:107] "Deleted CPUSet assignment" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="extract-content" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.863558 4867 memory_manager.go:354] "RemoveStaleState removing state" podUID="700228d2-601e-4f5a-be91-d36f5ac39ab5" containerName="registry-server" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.864895 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.865913 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.971530 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.971819 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:26 crc kubenswrapper[4867]: I1212 09:46:26.972000 4867 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thws8\" (UniqueName: \"kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.074985 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.075053 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.075103 4867 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thws8\" (UniqueName: \"kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.075647 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.075653 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.106332 4867 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thws8\" (UniqueName: \"kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8\") pod \"community-operators-j9wjk\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.216901 4867 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:27 crc kubenswrapper[4867]: I1212 09:46:27.905744 4867 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:28 crc kubenswrapper[4867]: I1212 09:46:28.082915 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerStarted","Data":"859c16c121104b68e68671bb18c47c5689dd80200d5d5055508883da1debc4f2"} Dec 12 09:46:29 crc kubenswrapper[4867]: I1212 09:46:29.093909 4867 generic.go:334] "Generic (PLEG): container finished" podID="e001fa3b-0599-4ec8-980b-efbcae829009" containerID="78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064" exitCode=0 Dec 12 09:46:29 crc kubenswrapper[4867]: I1212 09:46:29.093970 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerDied","Data":"78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064"} Dec 12 09:46:30 crc kubenswrapper[4867]: I1212 09:46:30.103744 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerStarted","Data":"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a"} Dec 12 09:46:31 crc kubenswrapper[4867]: I1212 09:46:31.113959 4867 generic.go:334] "Generic (PLEG): container finished" podID="e001fa3b-0599-4ec8-980b-efbcae829009" containerID="aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a" exitCode=0 Dec 12 09:46:31 crc kubenswrapper[4867]: I1212 09:46:31.114018 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerDied","Data":"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a"} Dec 12 09:46:32 crc kubenswrapper[4867]: I1212 09:46:32.127944 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerStarted","Data":"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34"} Dec 12 09:46:32 crc kubenswrapper[4867]: I1212 09:46:32.151025 4867 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9wjk" podStartSLOduration=3.645705699 podStartE2EDuration="6.151007351s" podCreationTimestamp="2025-12-12 09:46:26 +0000 UTC" firstStartedPulling="2025-12-12 09:46:29.096066023 +0000 UTC m=+10676.667447292" lastFinishedPulling="2025-12-12 09:46:31.601367675 +0000 UTC m=+10679.172748944" observedRunningTime="2025-12-12 09:46:32.146787628 +0000 UTC m=+10679.718168907" watchObservedRunningTime="2025-12-12 09:46:32.151007351 +0000 UTC m=+10679.722388620" Dec 12 09:46:37 crc kubenswrapper[4867]: I1212 09:46:37.217177 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:37 crc kubenswrapper[4867]: I1212 09:46:37.217736 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:37 crc kubenswrapper[4867]: I1212 09:46:37.263377 4867 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:38 crc kubenswrapper[4867]: I1212 09:46:38.234580 4867 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:38 crc kubenswrapper[4867]: I1212 09:46:38.281519 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.201664 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j9wjk" podUID="e001fa3b-0599-4ec8-980b-efbcae829009" containerName="registry-server" containerID="cri-o://b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34" gracePeriod=2 Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.888672 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.996291 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities\") pod \"e001fa3b-0599-4ec8-980b-efbcae829009\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.996365 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thws8\" (UniqueName: \"kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8\") pod \"e001fa3b-0599-4ec8-980b-efbcae829009\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.996577 4867 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content\") pod \"e001fa3b-0599-4ec8-980b-efbcae829009\" (UID: \"e001fa3b-0599-4ec8-980b-efbcae829009\") " Dec 12 09:46:40 crc kubenswrapper[4867]: I1212 09:46:40.997804 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities" (OuterVolumeSpecName: "utilities") pod "e001fa3b-0599-4ec8-980b-efbcae829009" (UID: "e001fa3b-0599-4ec8-980b-efbcae829009"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.001631 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8" (OuterVolumeSpecName: "kube-api-access-thws8") pod "e001fa3b-0599-4ec8-980b-efbcae829009" (UID: "e001fa3b-0599-4ec8-980b-efbcae829009"). InnerVolumeSpecName "kube-api-access-thws8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.050001 4867 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e001fa3b-0599-4ec8-980b-efbcae829009" (UID: "e001fa3b-0599-4ec8-980b-efbcae829009"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.098834 4867 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-utilities\") on node \"crc\" DevicePath \"\"" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.098869 4867 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thws8\" (UniqueName: \"kubernetes.io/projected/e001fa3b-0599-4ec8-980b-efbcae829009-kube-api-access-thws8\") on node \"crc\" DevicePath \"\"" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.098879 4867 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e001fa3b-0599-4ec8-980b-efbcae829009-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.213683 4867 generic.go:334] "Generic (PLEG): container finished" podID="e001fa3b-0599-4ec8-980b-efbcae829009" containerID="b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34" exitCode=0 Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.213721 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerDied","Data":"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34"} Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.213746 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wjk" event={"ID":"e001fa3b-0599-4ec8-980b-efbcae829009","Type":"ContainerDied","Data":"859c16c121104b68e68671bb18c47c5689dd80200d5d5055508883da1debc4f2"} Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.213762 4867 scope.go:117] "RemoveContainer" containerID="b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.213772 4867 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wjk" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.247619 4867 scope.go:117] "RemoveContainer" containerID="aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.254345 4867 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.266772 4867 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j9wjk"] Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.274663 4867 scope.go:117] "RemoveContainer" containerID="78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.314511 4867 scope.go:117] "RemoveContainer" containerID="b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34" Dec 12 09:46:41 crc kubenswrapper[4867]: E1212 09:46:41.315135 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34\": container with ID starting with b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34 not found: ID does not exist" containerID="b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.315189 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34"} err="failed to get container status \"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34\": rpc error: code = NotFound desc = could not find container \"b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34\": container with ID starting with b1140613f4609245ab8ddf8f2f974c94f6f72cd10dcd7abe5041dec60f896d34 not found: ID does not exist" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.315240 4867 scope.go:117] "RemoveContainer" containerID="aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a" Dec 12 09:46:41 crc kubenswrapper[4867]: E1212 09:46:41.315690 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a\": container with ID starting with aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a not found: ID does not exist" containerID="aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.315740 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a"} err="failed to get container status \"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a\": rpc error: code = NotFound desc = could not find container \"aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a\": container with ID starting with aaef0c5d5179c9c8ff491e4ac15749ddbcadcdeae7bec78d85212ccdf2458e4a not found: ID does not exist" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.315770 4867 scope.go:117] "RemoveContainer" containerID="78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064" Dec 12 09:46:41 crc kubenswrapper[4867]: E1212 09:46:41.316126 4867 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064\": container with ID starting with 78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064 not found: ID does not exist" containerID="78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064" Dec 12 09:46:41 crc kubenswrapper[4867]: I1212 09:46:41.316159 4867 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064"} err="failed to get container status \"78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064\": rpc error: code = NotFound desc = could not find container \"78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064\": container with ID starting with 78c13797c1d8fe22b7d4ed371731f386d585de09ed9cb587e012f38d5f88c064 not found: ID does not exist" Dec 12 09:46:42 crc kubenswrapper[4867]: I1212 09:46:42.857154 4867 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e001fa3b-0599-4ec8-980b-efbcae829009" path="/var/lib/kubelet/pods/e001fa3b-0599-4ec8-980b-efbcae829009/volumes" Dec 12 09:47:58 crc kubenswrapper[4867]: I1212 09:47:58.989444 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:47:58 crc kubenswrapper[4867]: I1212 09:47:58.990092 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:48:28 crc kubenswrapper[4867]: I1212 09:48:28.989043 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:48:28 crc kubenswrapper[4867]: I1212 09:48:28.989710 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:48:58 crc kubenswrapper[4867]: I1212 09:48:58.989051 4867 patch_prober.go:28] interesting pod/machine-config-daemon-q8lbh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 12 09:48:58 crc kubenswrapper[4867]: I1212 09:48:58.989619 4867 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 12 09:48:58 crc kubenswrapper[4867]: I1212 09:48:58.989664 4867 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" Dec 12 09:48:58 crc kubenswrapper[4867]: I1212 09:48:58.990458 4867 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9486a6274c2ad06071243c6ef7c9e72e8c56f8224e3124bdf4595856442807f"} pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 12 09:48:58 crc kubenswrapper[4867]: I1212 09:48:58.990513 4867 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" podUID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerName="machine-config-daemon" containerID="cri-o://f9486a6274c2ad06071243c6ef7c9e72e8c56f8224e3124bdf4595856442807f" gracePeriod=600 Dec 12 09:48:59 crc kubenswrapper[4867]: I1212 09:48:59.940311 4867 generic.go:334] "Generic (PLEG): container finished" podID="f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae" containerID="f9486a6274c2ad06071243c6ef7c9e72e8c56f8224e3124bdf4595856442807f" exitCode=0 Dec 12 09:48:59 crc kubenswrapper[4867]: I1212 09:48:59.940398 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerDied","Data":"f9486a6274c2ad06071243c6ef7c9e72e8c56f8224e3124bdf4595856442807f"} Dec 12 09:48:59 crc kubenswrapper[4867]: I1212 09:48:59.940981 4867 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-q8lbh" event={"ID":"f99b86bf-5b7d-4b8a-ab9b-df4cd408ccae","Type":"ContainerStarted","Data":"58dbd1c8a504a603f78f05c5a863883e633c0d4cae15c8cc03204c728ad159e4"} Dec 12 09:48:59 crc kubenswrapper[4867]: I1212 09:48:59.941047 4867 scope.go:117] "RemoveContainer" containerID="294e528285bbf364f56876ac0c2e2742f02f7d6577762815c3cc7be8b323145a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116762441024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116762442017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116734760016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116734760015467 5ustar corecore